var/home/core/zuul-output/0000755000175000017500000000000015113755732014536 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114000052015455 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005467276415114000042017705 0ustar rootrootDec 03 06:50:06 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 06:50:06 crc restorecon[4820]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:06 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:50:07 crc restorecon[4820]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 06:50:07 crc kubenswrapper[4946]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 06:50:07 crc kubenswrapper[4946]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 06:50:07 crc kubenswrapper[4946]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 06:50:07 crc kubenswrapper[4946]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 06:50:07 crc kubenswrapper[4946]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 06:50:07 crc kubenswrapper[4946]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.368835 4946 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372383 4946 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372403 4946 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372408 4946 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372429 4946 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372435 4946 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372440 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372445 4946 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372451 4946 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372455 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372460 4946 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372465 4946 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372469 4946 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372473 4946 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372477 4946 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372481 4946 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372485 4946 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372489 4946 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372494 4946 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372498 4946 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372503 4946 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372506 4946 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372510 4946 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372514 4946 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372518 4946 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372522 4946 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372526 4946 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372530 4946 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372534 4946 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372539 4946 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372543 4946 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372547 4946 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372551 4946 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372554 4946 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372559 4946 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372563 4946 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372567 4946 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372571 4946 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372576 4946 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372587 4946 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372592 4946 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372596 4946 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372599 4946 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372603 4946 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372606 4946 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372610 4946 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372613 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372617 4946 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372620 4946 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372624 4946 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372627 4946 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372631 4946 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372634 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372638 4946 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372641 4946 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372645 4946 feature_gate.go:330] unrecognized feature gate: Example Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372648 4946 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372652 4946 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372655 4946 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372659 4946 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372665 4946 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372670 4946 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372673 4946 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372677 4946 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372681 4946 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372685 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372689 4946 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372695 4946 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372700 4946 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372705 4946 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372709 4946 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.372717 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373179 4946 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373194 4946 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373202 4946 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373208 4946 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373219 4946 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373225 4946 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373236 4946 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373242 4946 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373246 4946 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373250 4946 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373255 4946 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373259 4946 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373263 4946 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373267 4946 flags.go:64] FLAG: --cgroup-root="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373271 4946 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373275 4946 flags.go:64] FLAG: --client-ca-file="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373280 4946 flags.go:64] FLAG: --cloud-config="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373284 4946 flags.go:64] FLAG: --cloud-provider="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373288 4946 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373292 4946 flags.go:64] FLAG: --cluster-domain="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373296 4946 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373301 4946 flags.go:64] FLAG: --config-dir="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373305 4946 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373310 4946 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373316 4946 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373320 4946 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373324 4946 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373328 4946 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373333 4946 flags.go:64] FLAG: --contention-profiling="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373337 4946 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373341 4946 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373349 4946 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373353 4946 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373358 4946 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373363 4946 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373367 4946 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373371 4946 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373375 4946 flags.go:64] FLAG: --enable-server="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373379 4946 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373385 4946 flags.go:64] FLAG: --event-burst="100" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373389 4946 flags.go:64] FLAG: --event-qps="50" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373394 4946 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373398 4946 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373402 4946 flags.go:64] FLAG: --eviction-hard="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373407 4946 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373411 4946 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373415 4946 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373420 4946 flags.go:64] FLAG: --eviction-soft="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373424 4946 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373429 4946 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373433 4946 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373437 4946 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373441 4946 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373445 4946 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373449 4946 flags.go:64] FLAG: --feature-gates="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373454 4946 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373458 4946 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373463 4946 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373467 4946 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373472 4946 flags.go:64] FLAG: --healthz-port="10248" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373476 4946 flags.go:64] FLAG: --help="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373480 4946 flags.go:64] FLAG: --hostname-override="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373484 4946 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373491 4946 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373495 4946 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373499 4946 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373503 4946 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373507 4946 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373512 4946 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373516 4946 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373520 4946 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373524 4946 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373528 4946 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373532 4946 flags.go:64] FLAG: --kube-reserved="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373536 4946 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373540 4946 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373545 4946 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373549 4946 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373553 4946 flags.go:64] FLAG: --lock-file="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373557 4946 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373561 4946 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373565 4946 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373572 4946 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373576 4946 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373580 4946 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373584 4946 flags.go:64] FLAG: --logging-format="text" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373588 4946 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373592 4946 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373597 4946 flags.go:64] FLAG: --manifest-url="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373601 4946 flags.go:64] FLAG: --manifest-url-header="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373606 4946 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373611 4946 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373616 4946 flags.go:64] FLAG: --max-pods="110" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373620 4946 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373624 4946 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373631 4946 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373636 4946 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373640 4946 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373645 4946 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373649 4946 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373659 4946 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373664 4946 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373668 4946 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373672 4946 flags.go:64] FLAG: --pod-cidr="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373676 4946 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373683 4946 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373687 4946 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373691 4946 flags.go:64] FLAG: --pods-per-core="0" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373696 4946 flags.go:64] FLAG: --port="10250" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373700 4946 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373704 4946 flags.go:64] FLAG: --provider-id="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373709 4946 flags.go:64] FLAG: --qos-reserved="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373714 4946 flags.go:64] FLAG: --read-only-port="10255" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373718 4946 flags.go:64] FLAG: --register-node="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373722 4946 flags.go:64] FLAG: --register-schedulable="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373726 4946 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373734 4946 flags.go:64] FLAG: --registry-burst="10" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373753 4946 flags.go:64] FLAG: --registry-qps="5" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373758 4946 flags.go:64] FLAG: --reserved-cpus="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373762 4946 flags.go:64] FLAG: --reserved-memory="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373767 4946 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373771 4946 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373776 4946 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373780 4946 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373784 4946 flags.go:64] FLAG: --runonce="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373788 4946 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373792 4946 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373799 4946 flags.go:64] FLAG: --seccomp-default="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373808 4946 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373812 4946 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373817 4946 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373821 4946 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373825 4946 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373829 4946 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373833 4946 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373838 4946 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373842 4946 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373847 4946 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373852 4946 flags.go:64] FLAG: --system-cgroups="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373857 4946 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373865 4946 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373870 4946 flags.go:64] FLAG: --tls-cert-file="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373875 4946 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373880 4946 flags.go:64] FLAG: --tls-min-version="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373884 4946 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373889 4946 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373893 4946 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373897 4946 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373902 4946 flags.go:64] FLAG: --v="2" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373910 4946 flags.go:64] FLAG: --version="false" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373916 4946 flags.go:64] FLAG: --vmodule="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373924 4946 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.373928 4946 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374040 4946 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374045 4946 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374049 4946 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374054 4946 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374059 4946 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374062 4946 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374068 4946 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374074 4946 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374078 4946 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374082 4946 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374085 4946 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374090 4946 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374093 4946 feature_gate.go:330] unrecognized feature gate: Example Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374097 4946 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374100 4946 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374104 4946 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374107 4946 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374111 4946 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374115 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374118 4946 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374122 4946 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374125 4946 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374129 4946 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374134 4946 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374138 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374142 4946 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374147 4946 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374151 4946 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374154 4946 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374158 4946 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374161 4946 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374166 4946 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374170 4946 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374173 4946 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374177 4946 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374180 4946 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374184 4946 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374187 4946 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374194 4946 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374199 4946 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374203 4946 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374206 4946 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374210 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374213 4946 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374219 4946 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374223 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374227 4946 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374231 4946 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374234 4946 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374238 4946 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374243 4946 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374246 4946 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374250 4946 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374253 4946 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374257 4946 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374261 4946 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374265 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374270 4946 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374274 4946 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374278 4946 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374282 4946 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374286 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374290 4946 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374294 4946 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374298 4946 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374302 4946 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374305 4946 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374310 4946 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374313 4946 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374317 4946 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.374323 4946 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.374539 4946 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.415937 4946 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.416012 4946 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416164 4946 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416182 4946 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416193 4946 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416203 4946 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416212 4946 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416221 4946 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416229 4946 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416240 4946 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416252 4946 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416263 4946 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416273 4946 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416283 4946 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416292 4946 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416300 4946 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416308 4946 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416316 4946 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416323 4946 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416332 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416341 4946 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416348 4946 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416356 4946 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416364 4946 feature_gate.go:330] unrecognized feature gate: Example Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416375 4946 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416389 4946 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416397 4946 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416405 4946 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416415 4946 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416423 4946 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416433 4946 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416441 4946 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416452 4946 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416462 4946 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416473 4946 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416484 4946 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416493 4946 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416502 4946 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416511 4946 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416520 4946 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416528 4946 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416536 4946 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416548 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416558 4946 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416566 4946 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416574 4946 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416583 4946 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416591 4946 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416600 4946 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416610 4946 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416620 4946 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416630 4946 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416640 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416650 4946 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416660 4946 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416671 4946 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416681 4946 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416690 4946 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416699 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416708 4946 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416719 4946 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416730 4946 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416787 4946 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416802 4946 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416813 4946 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416823 4946 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416834 4946 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416844 4946 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416854 4946 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416865 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416878 4946 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416892 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.416904 4946 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.416923 4946 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417185 4946 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417202 4946 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417211 4946 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417220 4946 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417229 4946 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417243 4946 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417258 4946 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417271 4946 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417280 4946 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417288 4946 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417296 4946 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417304 4946 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417312 4946 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417321 4946 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417329 4946 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417336 4946 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417344 4946 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417352 4946 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417360 4946 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417369 4946 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417377 4946 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417386 4946 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417393 4946 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417404 4946 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417414 4946 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417423 4946 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417431 4946 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417439 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417451 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417462 4946 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417475 4946 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417485 4946 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417496 4946 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417506 4946 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417516 4946 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417527 4946 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417536 4946 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417549 4946 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417559 4946 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417569 4946 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417577 4946 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417585 4946 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417593 4946 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417601 4946 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417609 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417616 4946 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417624 4946 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417632 4946 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417640 4946 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417648 4946 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417656 4946 feature_gate.go:330] unrecognized feature gate: Example Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417664 4946 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417672 4946 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417679 4946 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417687 4946 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417695 4946 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417705 4946 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417715 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417723 4946 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417732 4946 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417778 4946 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417788 4946 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417797 4946 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417807 4946 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417817 4946 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417827 4946 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417835 4946 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417845 4946 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417853 4946 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417862 4946 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.417870 4946 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.417884 4946 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.418188 4946 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.423474 4946 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.423636 4946 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.424574 4946 server.go:997] "Starting client certificate rotation" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.424614 4946 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.425045 4946 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-12 04:01:44.695020945 +0000 UTC Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.425226 4946 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.443352 4946 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.445236 4946 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.445817 4946 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.467243 4946 log.go:25] "Validated CRI v1 runtime API" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.483691 4946 log.go:25] "Validated CRI v1 image API" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.486249 4946 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.490720 4946 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-06-40-53-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.490782 4946 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.508081 4946 manager.go:217] Machine: {Timestamp:2025-12-03 06:50:07.506483641 +0000 UTC m=+0.303173770 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:e734b8d2-0665-40ac-a46d-0333906fc43c BootID:53a21bb3-ddb6-4066-b3b5-69f07da5f7ca Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:3e:c0:a5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:3e:c0:a5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:8b:07:a7 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:97:df:36 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:f7:a7:13 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:8c:f9:47 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:ee:de:fb Speed:-1 Mtu:1496} {Name:ens7.44 MacAddress:52:54:00:a6:38:d9 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4e:bd:53:ee:13:2e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:66:83:1f:32:42:be Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.508340 4946 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.508530 4946 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.508929 4946 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.509104 4946 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.509136 4946 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.509379 4946 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.509390 4946 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.509605 4946 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.509648 4946 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.509911 4946 state_mem.go:36] "Initialized new in-memory state store" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.510454 4946 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.511132 4946 kubelet.go:418] "Attempting to sync node with API server" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.511157 4946 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.511184 4946 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.511200 4946 kubelet.go:324] "Adding apiserver pod source" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.511217 4946 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.513348 4946 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.513824 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.513949 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.513924 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.514037 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.514272 4946 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515213 4946 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515843 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515875 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515884 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515894 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515910 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515920 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515931 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515946 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515956 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.515966 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.516007 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.516018 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.516225 4946 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.516921 4946 server.go:1280] "Started kubelet" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.517598 4946 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.517631 4946 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.517560 4946 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.519013 4946 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.520290 4946 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.520342 4946 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.520400 4946 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 05:11:35.498174089 +0000 UTC Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.520522 4946 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.520539 4946 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 06:50:07 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.520605 4946 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.520649 4946 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.520890 4946 server.go:460] "Adding debug handlers to kubelet server" Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.521369 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.521442 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.521382 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.527360 4946 factory.go:55] Registering systemd factory Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.527390 4946 factory.go:221] Registration of the systemd container factory successfully Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.527859 4946 factory.go:153] Registering CRI-O factory Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.527997 4946 factory.go:221] Registration of the crio container factory successfully Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.528194 4946 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.528327 4946 factory.go:103] Registering Raw factory Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.528412 4946 manager.go:1196] Started watching for new ooms in manager Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.528097 4946 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187da1deda543d3a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 06:50:07.516867898 +0000 UTC m=+0.313558017,LastTimestamp:2025-12-03 06:50:07.516867898 +0000 UTC m=+0.313558017,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.530156 4946 manager.go:319] Starting recovery of all containers Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.545510 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.546278 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.546347 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.546358 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.546370 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.546777 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.546790 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.546800 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547583 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547607 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547618 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547630 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547642 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547667 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547678 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547689 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547699 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547709 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547721 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547733 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547759 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547770 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547781 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547793 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547804 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547814 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547828 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547840 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547852 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547862 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547903 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547915 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547925 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547935 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547945 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547955 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547965 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547975 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547985 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.547995 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548084 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548096 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548105 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548116 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548130 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548140 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548150 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548159 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548169 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548182 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548193 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548203 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548216 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548227 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548238 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548277 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548290 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548303 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548316 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548326 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548338 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548348 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548357 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548366 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548375 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548384 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548431 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548439 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548449 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548460 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548469 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548477 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548486 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548498 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548509 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548520 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548532 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548541 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548550 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548560 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548569 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548579 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548591 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548601 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548613 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548624 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548634 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548645 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548656 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548666 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548677 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548687 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548698 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548708 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548718 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548728 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548753 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548765 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548775 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548785 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548795 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548807 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548818 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548829 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548845 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548857 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548900 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548911 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548924 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548935 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548947 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548958 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548969 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548980 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.548990 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.549001 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.549013 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.549024 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.549036 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.549046 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.551198 4946 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.551221 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.551232 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.551243 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.551254 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.551944 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552029 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552065 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552098 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552133 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552165 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552193 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552226 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552257 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552286 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552332 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552361 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552409 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552441 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552471 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552502 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552530 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552556 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552585 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552617 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552647 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552674 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552702 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552729 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552811 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552840 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552886 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552919 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552947 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.552977 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553005 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553033 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553093 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553126 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553154 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553181 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553212 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553245 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553271 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553297 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553327 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553353 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553384 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553411 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553438 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553464 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553491 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553520 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553548 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553574 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553602 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553627 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553658 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553701 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553728 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553850 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553881 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553910 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553938 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553966 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.553994 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554019 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554047 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554079 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554171 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554207 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554234 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554260 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554312 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554337 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554361 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554388 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554413 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554436 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554458 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554482 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554505 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554528 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554552 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554577 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554603 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554630 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554655 4946 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554681 4946 reconstruct.go:97] "Volume reconstruction finished" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.554700 4946 reconciler.go:26] "Reconciler: start to sync state" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.555736 4946 manager.go:324] Recovery completed Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.567101 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.571922 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.572039 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.572066 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.575030 4946 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.575070 4946 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.575105 4946 state_mem.go:36] "Initialized new in-memory state store" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.587940 4946 policy_none.go:49] "None policy: Start" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.588632 4946 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.589140 4946 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.589183 4946 state_mem.go:35] "Initializing new in-memory state store" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.591389 4946 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.591448 4946 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.591485 4946 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.591556 4946 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 06:50:07 crc kubenswrapper[4946]: W1203 06:50:07.592331 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.592402 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.621449 4946 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.659198 4946 manager.go:334] "Starting Device Plugin manager" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.660251 4946 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.660302 4946 server.go:79] "Starting device plugin registration server" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.660956 4946 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.660987 4946 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.661580 4946 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.661712 4946 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.661734 4946 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.671715 4946 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.692442 4946 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.692567 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.694019 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.694104 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.694128 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.694468 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.694850 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.694957 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696072 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696131 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696151 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696171 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696206 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696234 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696395 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696485 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.696537 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.697830 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.697896 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.697923 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.698257 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.698306 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.698324 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.698496 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.698686 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.698773 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.699676 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.699718 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.699758 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.699935 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.699964 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.699977 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.700002 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.700214 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.700281 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.701263 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.701292 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.701348 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.701492 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.701528 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.702182 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.702247 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.702262 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.702483 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.702503 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.702514 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.727993 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756725 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756791 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756833 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756861 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756884 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756906 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756928 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756953 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756972 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.756989 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.757072 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.757166 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.757214 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.757255 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.757295 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.761296 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.762663 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.762847 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.762874 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.762894 4946 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.763380 4946 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.858896 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.858981 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859016 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859047 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859083 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859158 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859213 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859259 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859261 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859304 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859321 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859352 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859305 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859273 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859282 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859391 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859306 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859341 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859602 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859696 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859704 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859833 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859762 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859865 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859889 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859899 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859909 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859931 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.859933 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.860054 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.963488 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.965400 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.965464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.965484 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:07 crc kubenswrapper[4946]: I1203 06:50:07.965527 4946 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:50:07 crc kubenswrapper[4946]: E1203 06:50:07.966125 4946 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.029132 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.040060 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.062207 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.069765 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e29cd0b2f754c7bdbc624f1cf3b0724ffe83f05cfac770dc3b63fd6583ecf306 WatchSource:0}: Error finding container e29cd0b2f754c7bdbc624f1cf3b0724ffe83f05cfac770dc3b63fd6583ecf306: Status 404 returned error can't find the container with id e29cd0b2f754c7bdbc624f1cf3b0724ffe83f05cfac770dc3b63fd6583ecf306 Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.074807 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-0471ef4606fcfec6de8787a41d754caa96f45ad3d49b924d3a6bb54b037bed43 WatchSource:0}: Error finding container 0471ef4606fcfec6de8787a41d754caa96f45ad3d49b924d3a6bb54b037bed43: Status 404 returned error can't find the container with id 0471ef4606fcfec6de8787a41d754caa96f45ad3d49b924d3a6bb54b037bed43 Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.087590 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.087815 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-bf3bded8d9a53eec338af51d24efd90cc258dadfdbc7fcfefd1cf08a99a68a54 WatchSource:0}: Error finding container bf3bded8d9a53eec338af51d24efd90cc258dadfdbc7fcfefd1cf08a99a68a54: Status 404 returned error can't find the container with id bf3bded8d9a53eec338af51d24efd90cc258dadfdbc7fcfefd1cf08a99a68a54 Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.092852 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.118395 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-75176d8c914d25f0122f7d3b663faf921390ab1682f152d41e3897d5d8f294c1 WatchSource:0}: Error finding container 75176d8c914d25f0122f7d3b663faf921390ab1682f152d41e3897d5d8f294c1: Status 404 returned error can't find the container with id 75176d8c914d25f0122f7d3b663faf921390ab1682f152d41e3897d5d8f294c1 Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.119183 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-eb0ebe33765de147437c0cf2a5717d841f52409bf673b3d59bb6ea22bbc50836 WatchSource:0}: Error finding container eb0ebe33765de147437c0cf2a5717d841f52409bf673b3d59bb6ea22bbc50836: Status 404 returned error can't find the container with id eb0ebe33765de147437c0cf2a5717d841f52409bf673b3d59bb6ea22bbc50836 Dec 03 06:50:08 crc kubenswrapper[4946]: E1203 06:50:08.129625 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.367228 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.368571 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.368611 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.368622 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.368650 4946 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:50:08 crc kubenswrapper[4946]: E1203 06:50:08.369011 4946 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.390768 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:08 crc kubenswrapper[4946]: E1203 06:50:08.390919 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.451113 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:08 crc kubenswrapper[4946]: E1203 06:50:08.451684 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.518562 4946 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.520974 4946 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 01:00:10.136376581 +0000 UTC Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.521046 4946 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 978h10m1.61533269s for next certificate rotation Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.599116 4946 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd" exitCode=0 Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.599220 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.599367 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e29cd0b2f754c7bdbc624f1cf3b0724ffe83f05cfac770dc3b63fd6583ecf306"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.599486 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.601236 4946 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f" exitCode=0 Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.601288 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.601315 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.601323 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.601342 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.601401 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eb0ebe33765de147437c0cf2a5717d841f52409bf673b3d59bb6ea22bbc50836"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.601503 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.602220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.602248 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.602257 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.603402 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.603434 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"75176d8c914d25f0122f7d3b663faf921390ab1682f152d41e3897d5d8f294c1"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.604168 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.605513 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.605543 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.605557 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.605654 4946 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3" exitCode=0 Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.605792 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.605921 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bf3bded8d9a53eec338af51d24efd90cc258dadfdbc7fcfefd1cf08a99a68a54"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.606024 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.606641 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.606691 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.606702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.607805 4946 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186" exitCode=0 Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.607855 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.607888 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0471ef4606fcfec6de8787a41d754caa96f45ad3d49b924d3a6bb54b037bed43"} Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.607965 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.608822 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.608874 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:08 crc kubenswrapper[4946]: I1203 06:50:08.608890 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.615904 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:08 crc kubenswrapper[4946]: E1203 06:50:08.616021 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:08 crc kubenswrapper[4946]: W1203 06:50:08.855405 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:08 crc kubenswrapper[4946]: E1203 06:50:08.855493 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:08 crc kubenswrapper[4946]: E1203 06:50:08.930881 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.169153 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.170509 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.170570 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.170583 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.170623 4946 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:50:09 crc kubenswrapper[4946]: E1203 06:50:09.171536 4946 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.464770 4946 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 06:50:09 crc kubenswrapper[4946]: E1203 06:50:09.465783 4946 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.527989 4946 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.614250 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.614294 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.614305 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.614418 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.615648 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.615707 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.615727 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.616107 4946 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8" exitCode=0 Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.616188 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.616217 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.617017 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.617043 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.617051 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.620507 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.620527 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.620537 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.620613 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.621439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.621467 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.621478 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.637718 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.637904 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.639062 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.639179 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.639257 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.643460 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.643616 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.643802 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313"} Dec 03 06:50:09 crc kubenswrapper[4946]: I1203 06:50:09.643915 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d"} Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.650444 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4"} Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.650586 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.652147 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.652197 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.652221 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.654539 4946 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a" exitCode=0 Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.654598 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a"} Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.654694 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.654704 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.656246 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.656289 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.656306 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.656491 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.656530 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.656552 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.772247 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.774355 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.774421 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.774447 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.774488 4946 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.974530 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:10 crc kubenswrapper[4946]: I1203 06:50:10.993971 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.002283 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.661370 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db"} Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.661457 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4"} Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.661479 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d"} Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.661508 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.662417 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.662496 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.662921 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.662980 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.663012 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.663876 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.663944 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:11 crc kubenswrapper[4946]: I1203 06:50:11.663968 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.037087 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.675891 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969"} Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.675944 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.675967 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3"} Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.676023 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.676046 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.676117 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.676733 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.677943 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.678081 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.678163 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.678100 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.678354 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.678377 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.678477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.678536 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.678556 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:12 crc kubenswrapper[4946]: I1203 06:50:12.784861 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.481950 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.610069 4946 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.678847 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.678888 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.678895 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.678908 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.680684 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.680724 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.680760 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.680774 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.680813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.680832 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.681174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.681207 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:13 crc kubenswrapper[4946]: I1203 06:50:13.681220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.033824 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.681841 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.681967 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.682954 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.682985 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.682995 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.683798 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.683852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:14 crc kubenswrapper[4946]: I1203 06:50:14.683869 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:15 crc kubenswrapper[4946]: I1203 06:50:15.896477 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:15 crc kubenswrapper[4946]: I1203 06:50:15.896719 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:15 crc kubenswrapper[4946]: I1203 06:50:15.898228 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:15 crc kubenswrapper[4946]: I1203 06:50:15.898331 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:15 crc kubenswrapper[4946]: I1203 06:50:15.898459 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:16 crc kubenswrapper[4946]: I1203 06:50:16.914387 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:16 crc kubenswrapper[4946]: I1203 06:50:16.914677 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:16 crc kubenswrapper[4946]: I1203 06:50:16.916570 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:16 crc kubenswrapper[4946]: I1203 06:50:16.916671 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:16 crc kubenswrapper[4946]: I1203 06:50:16.916691 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:17 crc kubenswrapper[4946]: E1203 06:50:17.671853 4946 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 06:50:18 crc kubenswrapper[4946]: I1203 06:50:18.307954 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 06:50:18 crc kubenswrapper[4946]: I1203 06:50:18.309092 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:18 crc kubenswrapper[4946]: I1203 06:50:18.320953 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:18 crc kubenswrapper[4946]: I1203 06:50:18.321020 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:18 crc kubenswrapper[4946]: I1203 06:50:18.321044 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.181078 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.181358 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.183001 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.183071 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.183096 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.187986 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.320975 4946 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.321934 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.695486 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.696426 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.696484 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:19 crc kubenswrapper[4946]: I1203 06:50:19.696500 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:20 crc kubenswrapper[4946]: W1203 06:50:20.106382 4946 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 03 06:50:20 crc kubenswrapper[4946]: I1203 06:50:20.106535 4946 trace.go:236] Trace[492727325]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 06:50:10.104) (total time: 10001ms): Dec 03 06:50:20 crc kubenswrapper[4946]: Trace[492727325]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:50:20.106) Dec 03 06:50:20 crc kubenswrapper[4946]: Trace[492727325]: [10.001836176s] [10.001836176s] END Dec 03 06:50:20 crc kubenswrapper[4946]: E1203 06:50:20.106573 4946 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 06:50:20 crc kubenswrapper[4946]: I1203 06:50:20.520899 4946 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 06:50:20 crc kubenswrapper[4946]: E1203 06:50:20.532509 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 03 06:50:20 crc kubenswrapper[4946]: I1203 06:50:20.558654 4946 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 06:50:20 crc kubenswrapper[4946]: I1203 06:50:20.558733 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 06:50:20 crc kubenswrapper[4946]: I1203 06:50:20.564072 4946 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 06:50:20 crc kubenswrapper[4946]: I1203 06:50:20.564159 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.044887 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.045215 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.047464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.047535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.047554 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.049489 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.181967 4946 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.182089 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.707000 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.708719 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.708938 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:22 crc kubenswrapper[4946]: I1203 06:50:22.709086 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.551586 4946 trace.go:236] Trace[659033472]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 06:50:10.762) (total time: 14789ms): Dec 03 06:50:25 crc kubenswrapper[4946]: Trace[659033472]: ---"Objects listed" error: 14789ms (06:50:25.551) Dec 03 06:50:25 crc kubenswrapper[4946]: Trace[659033472]: [14.789173379s] [14.789173379s] END Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.551689 4946 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.551795 4946 trace.go:236] Trace[2054928191]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 06:50:11.449) (total time: 14101ms): Dec 03 06:50:25 crc kubenswrapper[4946]: Trace[2054928191]: ---"Objects listed" error: 14101ms (06:50:25.551) Dec 03 06:50:25 crc kubenswrapper[4946]: Trace[2054928191]: [14.101716095s] [14.101716095s] END Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.551829 4946 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.552446 4946 trace.go:236] Trace[543977860]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 06:50:10.628) (total time: 14924ms): Dec 03 06:50:25 crc kubenswrapper[4946]: Trace[543977860]: ---"Objects listed" error: 14924ms (06:50:25.552) Dec 03 06:50:25 crc kubenswrapper[4946]: Trace[543977860]: [14.924331843s] [14.924331843s] END Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.552466 4946 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.552469 4946 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 06:50:25 crc kubenswrapper[4946]: E1203 06:50:25.556232 4946 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.582816 4946 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34094->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.582888 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34094->192.168.126.11:17697: read: connection reset by peer" Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.583246 4946 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.583294 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.583712 4946 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.583837 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.585317 4946 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.718471 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.721725 4946 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4" exitCode=255 Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.721794 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4"} Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.829669 4946 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 06:50:25 crc kubenswrapper[4946]: I1203 06:50:25.830241 4946 scope.go:117] "RemoveContainer" containerID="7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.524295 4946 apiserver.go:52] "Watching apiserver" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.529514 4946 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.529934 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.530279 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.530401 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.530547 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.530620 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.530790 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.530541 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.530917 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.531212 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.531292 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.533851 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.534929 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.535374 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.535755 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.537473 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.537527 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.537726 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.538185 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.538411 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.609078 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.621637 4946 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.624151 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.637793 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.653529 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.660549 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.660813 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.660942 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.661089 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.661233 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.661375 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.661544 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.661686 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.661876 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662024 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662167 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662295 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662423 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.661462 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.661604 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662003 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662065 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662563 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662657 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662692 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662719 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662104 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662769 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662796 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662821 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662843 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662864 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662885 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662903 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662926 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662946 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662971 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662223 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662260 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662318 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662545 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662932 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662972 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663156 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663228 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663297 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663330 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663352 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663380 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.662995 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663458 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663508 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663553 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663581 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663601 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663631 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663660 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663686 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663710 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663769 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663768 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663771 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663817 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663824 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663854 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663870 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663877 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663899 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663920 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663916 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663938 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.663945 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664049 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664069 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664074 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664130 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664184 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664234 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664287 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664335 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664385 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664433 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664518 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664090 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664556 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664555 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664573 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664150 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664140 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664198 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664236 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664627 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664677 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664727 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664814 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664861 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664906 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664951 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664998 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665137 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665196 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665242 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665287 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665335 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665381 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665474 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665522 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665570 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665620 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665670 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665713 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665836 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665929 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665972 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666013 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666057 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666100 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666143 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666188 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666234 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666287 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666332 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666377 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666420 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666470 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666551 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666589 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666621 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666657 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666694 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666727 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667137 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667183 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667235 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664338 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664355 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664534 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664534 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664700 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664718 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664757 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664860 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664882 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.664930 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665004 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665066 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665238 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665241 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665313 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665304 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667711 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667778 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665400 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665444 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665449 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667810 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667861 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667910 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667956 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667987 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668017 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668049 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668079 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668109 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668141 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668174 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668207 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668237 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668270 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668428 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668512 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668565 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668614 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668664 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668715 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668801 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668854 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668903 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668981 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669016 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669049 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669086 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669116 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669149 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669184 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669215 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669249 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669285 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669320 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669355 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669389 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669420 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669454 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669485 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669518 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669552 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669583 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669613 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669645 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669676 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669707 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669764 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669800 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669834 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669867 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669898 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669930 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669961 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669994 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670030 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670062 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670096 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670130 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670163 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670194 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670227 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670262 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670294 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670329 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670369 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670403 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670436 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670471 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670505 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670539 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670526 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670570 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670630 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670666 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670702 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665522 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665555 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671106 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665607 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665753 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671150 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665771 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.665733 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666089 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666110 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666317 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666391 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666712 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666785 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.666991 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667444 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.667498 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668137 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668269 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668435 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668508 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668580 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668788 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.668947 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669005 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.669331 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670048 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670214 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670721 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670734 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670703 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670793 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670821 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671040 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671166 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671403 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671490 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671551 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671693 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671692 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.671701 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.671708 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:50:27.17168878 +0000 UTC m=+19.968378879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672022 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672174 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672310 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672308 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672391 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672467 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672465 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672775 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.672031 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.670735 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673154 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673192 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673194 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673227 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673265 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673298 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673323 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673347 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673383 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673419 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673458 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673485 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673507 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673531 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673556 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673598 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673624 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673653 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673685 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673711 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673760 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673763 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673792 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673826 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673858 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.673950 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674018 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674021 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674029 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674060 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674099 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674126 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674150 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674174 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674195 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674218 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674249 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674276 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674300 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674311 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674327 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674343 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674355 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674380 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674526 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674545 4946 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674560 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674574 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674587 4946 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674599 4946 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674612 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674624 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674638 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674650 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674663 4946 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674678 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674694 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674707 4946 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674720 4946 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674733 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674785 4946 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674796 4946 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674810 4946 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674826 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674839 4946 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674852 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674865 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674877 4946 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674889 4946 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674902 4946 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674914 4946 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674926 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674939 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674951 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674965 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674980 4946 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.674993 4946 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675006 4946 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675020 4946 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675034 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675009 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675047 4946 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675063 4946 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675078 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675093 4946 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675106 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675122 4946 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675135 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675149 4946 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675161 4946 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675173 4946 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675185 4946 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675197 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675212 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675225 4946 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675239 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675254 4946 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675267 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675279 4946 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675291 4946 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675302 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675314 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675328 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.719916 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.719942 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.719976 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.719998 4946 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720015 4946 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720042 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720061 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720078 4946 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720093 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720118 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720134 4946 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720149 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720166 4946 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720189 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720206 4946 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720220 4946 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720236 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720258 4946 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720274 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720293 4946 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720316 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720331 4946 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720347 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720362 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720382 4946 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720398 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720411 4946 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720425 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720482 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720498 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720511 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720530 4946 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720544 4946 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720559 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720574 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720600 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720616 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720631 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720647 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720667 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720681 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720697 4946 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720714 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720733 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720768 4946 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720783 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720802 4946 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720817 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720833 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720849 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720869 4946 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720884 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.722402 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675102 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675138 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675251 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675314 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675319 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675541 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675574 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675590 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.727127 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675825 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.675921 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.676078 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.676212 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.676313 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.676501 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.676521 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.676726 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.676818 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.677053 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.677472 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.677520 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.677937 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.678180 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.678233 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.678702 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.678871 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.678916 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.678890 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.679591 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.679894 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.680414 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.680549 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.717517 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.717590 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.718144 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.718324 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.718320 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.718413 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.718780 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.719008 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.719038 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.719111 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720009 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720201 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720297 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720328 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720686 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720883 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.721049 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.720015 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.721539 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.719480 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.721864 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.722433 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.727820 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.722477 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.722859 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.722968 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723029 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723266 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723262 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723346 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.727995 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723435 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723474 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723643 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723780 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.723889 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.724102 4946 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.724356 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.724899 4946 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.725039 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.725566 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.725658 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.725788 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.726512 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.728108 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.728366 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.728465 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:27.228416994 +0000 UTC m=+20.025107113 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.728512 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:27.228489246 +0000 UTC m=+20.025179355 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.728920 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.730137 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.732978 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.733857 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.734799 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.738633 4946 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.738882 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.739127 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.739514 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41"} Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.739794 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.739934 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.740684 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.740712 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.740728 4946 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.740815 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:27.240798056 +0000 UTC m=+20.037488175 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.743843 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.746700 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.748391 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.749428 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.749820 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.750115 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.750138 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.750151 4946 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:26 crc kubenswrapper[4946]: E1203 06:50:26.750203 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:27.25018489 +0000 UTC m=+20.046874999 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.750931 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.752819 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.758769 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.762331 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.764809 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.770387 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.774131 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.783817 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.793084 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.806508 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821337 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821612 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821668 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821765 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821772 4946 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821836 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821852 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821866 4946 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821880 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821881 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821896 4946 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821911 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821925 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821937 4946 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821950 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821963 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821976 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.821991 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822004 4946 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822016 4946 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822031 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822047 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822064 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822081 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822097 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822138 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822156 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822174 4946 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822191 4946 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822206 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822223 4946 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822245 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822262 4946 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822279 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822298 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822611 4946 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822637 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822777 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822809 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822846 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822861 4946 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822875 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822889 4946 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822902 4946 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822915 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822948 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822961 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822973 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822984 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.822996 4946 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823008 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823021 4946 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823032 4946 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823044 4946 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823056 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823090 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823103 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823118 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823131 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823145 4946 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823159 4946 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823178 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823191 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823207 4946 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823220 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823233 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823247 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823259 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823272 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823307 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823324 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823336 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823349 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823364 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823376 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823389 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823401 4946 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823412 4946 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823424 4946 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823435 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823448 4946 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823461 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823475 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823516 4946 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823530 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823546 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823558 4946 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823571 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.823583 4946 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.832273 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.854466 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:50:26 crc kubenswrapper[4946]: W1203 06:50:26.867560 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-03cf9b88fc0adae05efaf8b46f940b46a68c20b69b85cab2f3533a0c108cc09d WatchSource:0}: Error finding container 03cf9b88fc0adae05efaf8b46f940b46a68c20b69b85cab2f3533a0c108cc09d: Status 404 returned error can't find the container with id 03cf9b88fc0adae05efaf8b46f940b46a68c20b69b85cab2f3533a0c108cc09d Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.870565 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:50:26 crc kubenswrapper[4946]: I1203 06:50:26.882437 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:50:26 crc kubenswrapper[4946]: W1203 06:50:26.903509 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-c17652e5aecd1b104f0932861a8f722eb6ef8724a33fc615fad8ee55a4abf9ea WatchSource:0}: Error finding container c17652e5aecd1b104f0932861a8f722eb6ef8724a33fc615fad8ee55a4abf9ea: Status 404 returned error can't find the container with id c17652e5aecd1b104f0932861a8f722eb6ef8724a33fc615fad8ee55a4abf9ea Dec 03 06:50:26 crc kubenswrapper[4946]: W1203 06:50:26.907548 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-4dfdbafb4719f0f765938117e7431962cd88ae073b9308d2c5908cd773c7e1ea WatchSource:0}: Error finding container 4dfdbafb4719f0f765938117e7431962cd88ae073b9308d2c5908cd773c7e1ea: Status 404 returned error can't find the container with id 4dfdbafb4719f0f765938117e7431962cd88ae073b9308d2c5908cd773c7e1ea Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.226670 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.226849 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:50:28.226819036 +0000 UTC m=+21.023509155 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.327581 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.327632 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.327659 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.327691 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.327833 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.327852 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.327867 4946 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.327919 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:28.327903243 +0000 UTC m=+21.124593372 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.328014 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.328060 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.328026 4946 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.328172 4946 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.328197 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:28.32817379 +0000 UTC m=+21.124863919 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.328078 4946 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.328342 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:28.328295973 +0000 UTC m=+21.124986122 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.328376 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:28.328356645 +0000 UTC m=+21.125046764 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.592714 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:27 crc kubenswrapper[4946]: E1203 06:50:27.592997 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.597583 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.598674 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.600370 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.601654 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.602881 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.603886 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.605292 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.606612 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.609044 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.610402 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.611549 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.614101 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.615389 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.616323 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.616797 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.618087 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.619382 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.620878 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.621716 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.624289 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.625708 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.626721 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.628662 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.629619 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.631084 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.632022 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.633302 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.634734 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.634944 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.635729 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.637095 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.638177 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.639160 4946 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.639362 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.642471 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.643961 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.644393 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.646353 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.647699 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.648896 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.650682 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.652379 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.653491 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.655030 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.656829 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.658166 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.658480 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.659398 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.661025 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.662098 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.663691 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.664425 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.665654 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.666872 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.667913 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.668715 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.673543 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.680900 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.705564 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.730564 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.744898 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f"} Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.744976 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"03cf9b88fc0adae05efaf8b46f940b46a68c20b69b85cab2f3533a0c108cc09d"} Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.746813 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4dfdbafb4719f0f765938117e7431962cd88ae073b9308d2c5908cd773c7e1ea"} Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.750516 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0"} Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.750577 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f"} Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.750600 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c17652e5aecd1b104f0932861a8f722eb6ef8724a33fc615fad8ee55a4abf9ea"} Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.752793 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.771514 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.787141 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.808541 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.831166 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.850459 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.884191 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:27 crc kubenswrapper[4946]: I1203 06:50:27.900588 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.239307 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.239481 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:50:30.239459331 +0000 UTC m=+23.036149450 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.340954 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.341019 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.341051 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.341089 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341225 4946 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341244 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341295 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:30.341274457 +0000 UTC m=+23.137964586 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341290 4946 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341413 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341297 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341460 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341478 4946 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341485 4946 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341429 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:30.34140142 +0000 UTC m=+23.138091559 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341598 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:30.341569694 +0000 UTC m=+23.138259843 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.341627 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:30.341611455 +0000 UTC m=+23.138301604 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.342248 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.358066 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.359643 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.369911 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.387184 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.394563 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.402725 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.420392 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.438301 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.456066 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.471627 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.488958 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.512868 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.534817 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.548797 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.562425 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.577815 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.591857 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.591968 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.592030 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.592163 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.593545 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.756920 4946 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.758571 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.758613 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.758627 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.758715 4946 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.766310 4946 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.766602 4946 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.767949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.768009 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.768025 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.768044 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.768062 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:28Z","lastTransitionTime":"2025-12-03T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.788607 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.792437 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.792479 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.792491 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.792508 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.792520 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:28Z","lastTransitionTime":"2025-12-03T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.805018 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.809159 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.809207 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.809222 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.809242 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.809256 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:28Z","lastTransitionTime":"2025-12-03T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.823734 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.827798 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.827868 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.827888 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.827916 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.827936 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:28Z","lastTransitionTime":"2025-12-03T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.844959 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.849108 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.849153 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.849170 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.849196 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.849211 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:28Z","lastTransitionTime":"2025-12-03T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.862180 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:28 crc kubenswrapper[4946]: E1203 06:50:28.862341 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.864103 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.864138 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.864153 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.864169 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.864182 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:28Z","lastTransitionTime":"2025-12-03T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.967241 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.967310 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.967328 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.967356 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:28 crc kubenswrapper[4946]: I1203 06:50:28.967376 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:28Z","lastTransitionTime":"2025-12-03T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.069702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.069768 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.069782 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.069799 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.069811 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.173941 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.174007 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.174030 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.174059 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.174084 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.186315 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.193173 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.199836 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.220998 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.245503 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.267649 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.277682 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.277744 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.277801 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.277831 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.277855 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.290643 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.309615 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.335993 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.354478 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.370873 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.381633 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.381698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.381717 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.381750 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.381805 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.390562 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.413950 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.435446 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.454521 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.489207 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.489268 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.489285 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.489314 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.489333 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.495942 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.520602 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.540362 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.561893 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.579923 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.591768 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:29 crc kubenswrapper[4946]: E1203 06:50:29.591891 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.592055 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.592116 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.592134 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.592160 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.592177 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.695212 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.695274 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.695293 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.695319 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.695337 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: E1203 06:50:29.765056 4946 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.798208 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.798274 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.798294 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.798324 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.798344 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.901546 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.901626 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.901649 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.901680 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:29 crc kubenswrapper[4946]: I1203 06:50:29.901702 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:29Z","lastTransitionTime":"2025-12-03T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.005544 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.005639 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.005676 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.005706 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.005729 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.109016 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.109095 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.109117 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.109152 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.109175 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.211554 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.211592 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.211603 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.211618 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.211628 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.258072 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.258247 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:50:34.25821003 +0000 UTC m=+27.054900179 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.315037 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.315116 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.315139 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.315171 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.315193 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.359645 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.359718 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.359795 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.359837 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.359898 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.359947 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.359970 4946 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.359984 4946 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.360014 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.360049 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.360075 4946 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.360082 4946 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.360051 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:34.360023956 +0000 UTC m=+27.156714095 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.360179 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:34.360143969 +0000 UTC m=+27.156834118 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.360215 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:34.36019932 +0000 UTC m=+27.156889539 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.360244 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:34.360228721 +0000 UTC m=+27.156918930 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.418746 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.418858 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.418883 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.418914 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.418938 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.522700 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.522833 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.522864 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.522895 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.522919 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.592499 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.592499 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.592714 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:30 crc kubenswrapper[4946]: E1203 06:50:30.592869 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.625723 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.625811 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.625829 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.625853 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.625872 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.729639 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.729732 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.729813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.729848 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.729875 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.760048 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.783473 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.803070 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.821154 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.833677 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.833731 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.833766 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.833787 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.833801 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.835824 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.852693 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.870320 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.887024 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.912061 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.936827 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.936906 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.936925 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.936949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.936964 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:30Z","lastTransitionTime":"2025-12-03T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:30 crc kubenswrapper[4946]: I1203 06:50:30.939896 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.039297 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.039369 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.039388 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.039414 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.039432 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.141763 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.141797 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.141806 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.141821 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.141832 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.244950 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.245001 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.245020 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.245040 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.245055 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.347968 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.348028 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.348045 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.348069 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.348086 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.451481 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.451547 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.451564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.451591 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.451610 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.555480 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.555550 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.555576 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.555603 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.555621 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.592265 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:31 crc kubenswrapper[4946]: E1203 06:50:31.592506 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.657853 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.657908 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.657920 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.657940 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.657951 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.760438 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.760493 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.760508 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.760527 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.760541 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.863688 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.863772 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.863792 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.863815 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.863833 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.967398 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.967483 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.967503 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.967531 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:31 crc kubenswrapper[4946]: I1203 06:50:31.967549 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:31Z","lastTransitionTime":"2025-12-03T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.071165 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.071257 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.071270 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.071289 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.071302 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.173791 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.173850 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.173863 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.173885 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.173902 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.276527 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.276567 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.276578 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.276597 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.276611 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.339298 4946 csr.go:261] certificate signing request csr-rcd6l is approved, waiting to be issued Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.352587 4946 csr.go:257] certificate signing request csr-rcd6l is issued Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.379084 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.379132 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.379142 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.379160 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.379169 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.481800 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.481842 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.481850 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.481866 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.481876 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.586158 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.586211 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.586224 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.586256 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.586275 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.591774 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.591895 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:32 crc kubenswrapper[4946]: E1203 06:50:32.591935 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:32 crc kubenswrapper[4946]: E1203 06:50:32.592092 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.688877 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.688933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.688945 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.688963 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.688973 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.739235 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-2b4cb"] Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.739711 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2b4cb" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.742269 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.742251 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.742549 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.759288 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.775418 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.781332 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sgrs\" (UniqueName: \"kubernetes.io/projected/188b3f12-d66a-4447-979f-efea0e31abf1-kube-api-access-2sgrs\") pod \"node-resolver-2b4cb\" (UID: \"188b3f12-d66a-4447-979f-efea0e31abf1\") " pod="openshift-dns/node-resolver-2b4cb" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.781400 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/188b3f12-d66a-4447-979f-efea0e31abf1-hosts-file\") pod \"node-resolver-2b4cb\" (UID: \"188b3f12-d66a-4447-979f-efea0e31abf1\") " pod="openshift-dns/node-resolver-2b4cb" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.790026 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.791090 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.791127 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.791136 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.791151 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.791162 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.804729 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.816660 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.832141 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.845028 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.866819 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.880921 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.882191 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/188b3f12-d66a-4447-979f-efea0e31abf1-hosts-file\") pod \"node-resolver-2b4cb\" (UID: \"188b3f12-d66a-4447-979f-efea0e31abf1\") " pod="openshift-dns/node-resolver-2b4cb" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.882253 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sgrs\" (UniqueName: \"kubernetes.io/projected/188b3f12-d66a-4447-979f-efea0e31abf1-kube-api-access-2sgrs\") pod \"node-resolver-2b4cb\" (UID: \"188b3f12-d66a-4447-979f-efea0e31abf1\") " pod="openshift-dns/node-resolver-2b4cb" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.882402 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/188b3f12-d66a-4447-979f-efea0e31abf1-hosts-file\") pod \"node-resolver-2b4cb\" (UID: \"188b3f12-d66a-4447-979f-efea0e31abf1\") " pod="openshift-dns/node-resolver-2b4cb" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.893795 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.893845 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.893857 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.893877 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.893889 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.901439 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.901942 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sgrs\" (UniqueName: \"kubernetes.io/projected/188b3f12-d66a-4447-979f-efea0e31abf1-kube-api-access-2sgrs\") pod \"node-resolver-2b4cb\" (UID: \"188b3f12-d66a-4447-979f-efea0e31abf1\") " pod="openshift-dns/node-resolver-2b4cb" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.996946 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.997394 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.997469 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.997580 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:32 crc kubenswrapper[4946]: I1203 06:50:32.997667 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:32Z","lastTransitionTime":"2025-12-03T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.051674 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2b4cb" Dec 03 06:50:33 crc kubenswrapper[4946]: W1203 06:50:33.070298 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod188b3f12_d66a_4447_979f_efea0e31abf1.slice/crio-dedd4e816a5eb52b959dc5e6e8425cace846d5d656049dd5d4e4891837425870 WatchSource:0}: Error finding container dedd4e816a5eb52b959dc5e6e8425cace846d5d656049dd5d4e4891837425870: Status 404 returned error can't find the container with id dedd4e816a5eb52b959dc5e6e8425cace846d5d656049dd5d4e4891837425870 Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.100358 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.100639 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.100908 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.101057 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.101200 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.175396 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6fppr"] Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.176065 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.195233 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.195308 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.195248 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.195617 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.196104 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.196264 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.196383 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.196803 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-lzmlt"] Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.197450 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.207264 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.207292 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.207302 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.207318 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.207329 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.211616 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-t9hvz"] Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.212920 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-6bt2d"] Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.214104 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.214557 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.211358 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.221033 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.221119 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.221594 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.221041 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.232495 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.232554 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.232770 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.232795 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.232503 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.232980 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.233055 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.235064 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.272124 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286090 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-os-release\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286132 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-netns\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286151 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-node-log\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286186 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-ovn\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286203 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-netd\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286219 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-kubelet\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286235 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-ovn-kubernetes\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286254 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-script-lib\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286271 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-k8s-cni-cncf-io\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286289 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-conf-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286304 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-etc-kubernetes\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286329 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-var-lib-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286342 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/81237850-a445-4887-86e0-23bb0fa052c2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286359 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4003d158-6bdd-45bd-a68c-ca52bd7264c5-rootfs\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286376 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp9rm\" (UniqueName: \"kubernetes.io/projected/4003d158-6bdd-45bd-a68c-ca52bd7264c5-kube-api-access-pp9rm\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286391 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286405 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-netns\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286424 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-hostroot\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286445 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-env-overrides\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286459 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286476 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4003d158-6bdd-45bd-a68c-ca52bd7264c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286491 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-cnibin\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286508 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-daemon-config\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286522 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-systemd\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286535 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4003d158-6bdd-45bd-a68c-ca52bd7264c5-proxy-tls\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286551 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-cnibin\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286565 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovn-node-metrics-cert\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286581 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-systemd-units\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286596 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-system-cni-dir\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286610 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-system-cni-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286623 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/08074f18-fe84-4d7b-8327-9696cbe78f38-cni-binary-copy\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286638 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-multus-certs\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286654 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-cni-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286672 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-cni-multus\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286687 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286705 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-config\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286718 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-cni-bin\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286732 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d6hn\" (UniqueName: \"kubernetes.io/projected/08074f18-fe84-4d7b-8327-9696cbe78f38-kube-api-access-2d6hn\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286754 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-etc-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286782 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-log-socket\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286796 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-slash\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286811 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht6gh\" (UniqueName: \"kubernetes.io/projected/81237850-a445-4887-86e0-23bb0fa052c2-kube-api-access-ht6gh\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286826 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-socket-dir-parent\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286839 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-kubelet\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286855 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-bin\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286869 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzcsn\" (UniqueName: \"kubernetes.io/projected/58ed82b9-82e5-4d9a-b331-80c32397dc43-kube-api-access-lzcsn\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286897 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81237850-a445-4887-86e0-23bb0fa052c2-cni-binary-copy\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.286910 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-os-release\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.322971 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.323016 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.323054 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.323071 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.323083 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.329915 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.353784 4946 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-03 06:45:32 +0000 UTC, rotation deadline is 2026-10-04 04:32:56.453122562 +0000 UTC Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.353873 4946 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7317h42m23.099252637s for next certificate rotation Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.369808 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388297 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-kubelet\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388351 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-ovn-kubernetes\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388375 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-script-lib\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388398 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-k8s-cni-cncf-io\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388422 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-etc-kubernetes\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388444 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-conf-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388480 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-var-lib-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388485 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-ovn-kubernetes\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388500 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/81237850-a445-4887-86e0-23bb0fa052c2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388599 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-var-lib-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388476 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-kubelet\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388630 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-k8s-cni-cncf-io\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388584 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-etc-kubernetes\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388696 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-conf-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388790 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-env-overrides\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388816 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4003d158-6bdd-45bd-a68c-ca52bd7264c5-rootfs\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388839 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp9rm\" (UniqueName: \"kubernetes.io/projected/4003d158-6bdd-45bd-a68c-ca52bd7264c5-kube-api-access-pp9rm\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388859 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388851 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4003d158-6bdd-45bd-a68c-ca52bd7264c5-rootfs\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388875 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-netns\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388893 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-hostroot\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388912 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388931 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4003d158-6bdd-45bd-a68c-ca52bd7264c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388954 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-cnibin\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388974 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4003d158-6bdd-45bd-a68c-ca52bd7264c5-proxy-tls\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.388992 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-cnibin\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389008 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-daemon-config\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389024 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-systemd\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389041 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovn-node-metrics-cert\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389058 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-systemd-units\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389076 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-system-cni-dir\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389100 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-system-cni-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389120 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/08074f18-fe84-4d7b-8327-9696cbe78f38-cni-binary-copy\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389134 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-multus-certs\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389154 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-cni-multus\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389172 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-cni-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389189 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-etc-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389206 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-log-socket\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389225 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389241 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-systemd\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389245 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-config\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389297 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-cni-bin\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389322 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d6hn\" (UniqueName: \"kubernetes.io/projected/08074f18-fe84-4d7b-8327-9696cbe78f38-kube-api-access-2d6hn\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389362 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-slash\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389390 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-bin\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389417 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht6gh\" (UniqueName: \"kubernetes.io/projected/81237850-a445-4887-86e0-23bb0fa052c2-kube-api-access-ht6gh\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389442 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-socket-dir-parent\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389458 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-kubelet\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389483 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzcsn\" (UniqueName: \"kubernetes.io/projected/58ed82b9-82e5-4d9a-b331-80c32397dc43-kube-api-access-lzcsn\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389499 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81237850-a445-4887-86e0-23bb0fa052c2-cni-binary-copy\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389515 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-os-release\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389527 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-env-overrides\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389534 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/81237850-a445-4887-86e0-23bb0fa052c2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389555 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-netns\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389549 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-script-lib\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389587 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-multus-certs\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389594 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-slash\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389608 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-etc-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389625 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-cni-multus\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389632 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-bin\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389664 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-cnibin\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389841 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-socket-dir-parent\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389873 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-kubelet\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389911 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-config\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389928 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-cni-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389960 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-log-socket\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389990 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389995 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-run-netns\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390049 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-hostroot\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390066 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4003d158-6bdd-45bd-a68c-ca52bd7264c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390085 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-system-cni-dir\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390115 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-system-cni-dir\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390120 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-systemd-units\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390141 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-openvswitch\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390185 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-cnibin\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390186 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-os-release\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.389534 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-netns\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390245 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-node-log\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390270 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-os-release\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390297 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-ovn\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390319 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-netd\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390384 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-netd\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390419 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-node-log\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390464 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-os-release\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390495 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-ovn\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390535 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/08074f18-fe84-4d7b-8327-9696cbe78f38-host-var-lib-cni-bin\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390554 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81237850-a445-4887-86e0-23bb0fa052c2-cni-binary-copy\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390704 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/81237850-a445-4887-86e0-23bb0fa052c2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390748 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/08074f18-fe84-4d7b-8327-9696cbe78f38-multus-daemon-config\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.390805 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/08074f18-fe84-4d7b-8327-9696cbe78f38-cni-binary-copy\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.394146 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovn-node-metrics-cert\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.401275 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4003d158-6bdd-45bd-a68c-ca52bd7264c5-proxy-tls\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.405697 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.417440 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d6hn\" (UniqueName: \"kubernetes.io/projected/08074f18-fe84-4d7b-8327-9696cbe78f38-kube-api-access-2d6hn\") pod \"multus-t9hvz\" (UID: \"08074f18-fe84-4d7b-8327-9696cbe78f38\") " pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.419012 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht6gh\" (UniqueName: \"kubernetes.io/projected/81237850-a445-4887-86e0-23bb0fa052c2-kube-api-access-ht6gh\") pod \"multus-additional-cni-plugins-lzmlt\" (UID: \"81237850-a445-4887-86e0-23bb0fa052c2\") " pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.420430 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzcsn\" (UniqueName: \"kubernetes.io/projected/58ed82b9-82e5-4d9a-b331-80c32397dc43-kube-api-access-lzcsn\") pod \"ovnkube-node-6fppr\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.420455 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp9rm\" (UniqueName: \"kubernetes.io/projected/4003d158-6bdd-45bd-a68c-ca52bd7264c5-kube-api-access-pp9rm\") pod \"machine-config-daemon-6bt2d\" (UID: \"4003d158-6bdd-45bd-a68c-ca52bd7264c5\") " pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.422127 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.427579 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.427607 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.427616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.427631 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.427643 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.438842 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.460855 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.488114 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.488247 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: W1203 06:50:33.501229 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58ed82b9_82e5_4d9a_b331_80c32397dc43.slice/crio-25638cd36b4f7f640ad6d77aa55eeb01b44bddc032d173b438aa5bd72d527b63 WatchSource:0}: Error finding container 25638cd36b4f7f640ad6d77aa55eeb01b44bddc032d173b438aa5bd72d527b63: Status 404 returned error can't find the container with id 25638cd36b4f7f640ad6d77aa55eeb01b44bddc032d173b438aa5bd72d527b63 Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.507282 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.508915 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" Dec 03 06:50:33 crc kubenswrapper[4946]: W1203 06:50:33.522917 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81237850_a445_4887_86e0_23bb0fa052c2.slice/crio-ecd49ded8de07b1ec958202011f31da364f734e5c4993e8f0662ebf630f74369 WatchSource:0}: Error finding container ecd49ded8de07b1ec958202011f31da364f734e5c4993e8f0662ebf630f74369: Status 404 returned error can't find the container with id ecd49ded8de07b1ec958202011f31da364f734e5c4993e8f0662ebf630f74369 Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.523889 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.530639 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.530688 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.530705 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.530722 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.530737 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.538841 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.543737 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.556223 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-t9hvz" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.556292 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: W1203 06:50:33.558197 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4003d158_6bdd_45bd_a68c_ca52bd7264c5.slice/crio-81da883fea6678201779bb29426ef4a80094c5e02cbadd13172f4cb243ae374b WatchSource:0}: Error finding container 81da883fea6678201779bb29426ef4a80094c5e02cbadd13172f4cb243ae374b: Status 404 returned error can't find the container with id 81da883fea6678201779bb29426ef4a80094c5e02cbadd13172f4cb243ae374b Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.569332 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.587618 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.593472 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:33 crc kubenswrapper[4946]: E1203 06:50:33.593593 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.599062 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.612199 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.630110 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.634688 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.634733 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.634743 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.634775 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.634786 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.644316 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.657932 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.672896 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.685290 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.706332 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.720228 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.733075 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.736532 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.736567 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.736578 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.736595 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.736606 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.768492 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.768536 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"81da883fea6678201779bb29426ef4a80094c5e02cbadd13172f4cb243ae374b"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.770174 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerStarted","Data":"bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.770226 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerStarted","Data":"ecd49ded8de07b1ec958202011f31da364f734e5c4993e8f0662ebf630f74369"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.771495 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2b4cb" event={"ID":"188b3f12-d66a-4447-979f-efea0e31abf1","Type":"ContainerStarted","Data":"abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.771531 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2b4cb" event={"ID":"188b3f12-d66a-4447-979f-efea0e31abf1","Type":"ContainerStarted","Data":"dedd4e816a5eb52b959dc5e6e8425cace846d5d656049dd5d4e4891837425870"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.772917 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t9hvz" event={"ID":"08074f18-fe84-4d7b-8327-9696cbe78f38","Type":"ContainerStarted","Data":"79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.772944 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t9hvz" event={"ID":"08074f18-fe84-4d7b-8327-9696cbe78f38","Type":"ContainerStarted","Data":"b1d84c764425e48a3191a7db573e3e8f405baf5141c703ff73cdd5ba95f5134e"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.775137 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c" exitCode=0 Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.775175 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.775198 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"25638cd36b4f7f640ad6d77aa55eeb01b44bddc032d173b438aa5bd72d527b63"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.784286 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.794915 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.821229 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.833143 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.843117 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.843219 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.843234 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.843261 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.843273 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.859445 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.875184 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.887657 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.913611 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.937546 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.946012 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.946055 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.946067 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.946084 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.946097 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:33Z","lastTransitionTime":"2025-12-03T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.950807 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.975105 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:33 crc kubenswrapper[4946]: I1203 06:50:33.991085 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.004990 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.040646 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.048106 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.048159 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.048172 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.048212 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.048228 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.058705 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.085134 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.101682 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.117793 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.131948 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.150719 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.150784 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.150797 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.150814 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.150825 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.151409 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.172350 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.185912 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.205141 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.218313 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.231813 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.253274 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.253319 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.253332 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.253315 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.253351 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.253503 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.264973 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.282928 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.300075 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.300213 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:50:42.300194765 +0000 UTC m=+35.096884874 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.355772 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.355812 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.355824 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.355840 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.355850 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.401316 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.401377 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.401406 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.401439 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401505 4946 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401565 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401614 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:42.40158891 +0000 UTC m=+35.198279049 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401584 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401650 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401670 4946 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401716 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:42.401702773 +0000 UTC m=+35.198392912 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401616 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401770 4946 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401813 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:42.401801836 +0000 UTC m=+35.198491975 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401523 4946 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.401860 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:42.401848787 +0000 UTC m=+35.198538936 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.465764 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.465796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.465806 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.465821 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.465830 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.568449 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.568490 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.568499 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.568515 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.568526 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.592154 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.592709 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.592178 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:34 crc kubenswrapper[4946]: E1203 06:50:34.593170 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.670876 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.670911 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.670921 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.670938 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.670950 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.774311 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.774340 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.774350 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.774371 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.774409 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.780707 4946 generic.go:334] "Generic (PLEG): container finished" podID="81237850-a445-4887-86e0-23bb0fa052c2" containerID="bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15" exitCode=0 Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.780778 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerDied","Data":"bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.799518 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.800844 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.800896 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.800911 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.800925 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.800936 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.803324 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.812160 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.832246 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.843621 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.859971 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.872721 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.877939 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.878010 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.878024 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.878050 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.878060 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.897209 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.912500 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.934310 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.948366 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.962224 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.976214 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.982634 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.982665 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.982674 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.982689 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.982698 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:34Z","lastTransitionTime":"2025-12-03T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:34 crc kubenswrapper[4946]: I1203 06:50:34.989136 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.001383 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.021900 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.037900 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.052889 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.073357 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.090101 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.090149 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.090161 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.090182 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.090252 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.092686 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.109054 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.122939 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.133072 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.156470 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.170501 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.192522 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.192552 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.192560 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.192574 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.192582 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.199471 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.215133 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.232644 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.244875 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.294994 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.295036 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.295050 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.295068 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.295080 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.397420 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.397477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.397487 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.397505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.397516 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.426113 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-8plsb"] Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.426577 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.428415 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.428439 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.428857 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.429043 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.444299 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.454692 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.470477 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.485345 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.497992 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.499513 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.499566 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.499581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.499605 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.499621 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.513840 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b84d71a3-4e42-48a6-802a-553364d32a9b-host\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.513931 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b84d71a3-4e42-48a6-802a-553364d32a9b-serviceca\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.513971 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86mzf\" (UniqueName: \"kubernetes.io/projected/b84d71a3-4e42-48a6-802a-553364d32a9b-kube-api-access-86mzf\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.523121 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.539187 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.555683 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.569992 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.586853 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.591828 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:35 crc kubenswrapper[4946]: E1203 06:50:35.591979 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.602517 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.602521 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.602569 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.602682 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.602705 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.602717 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.614910 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b84d71a3-4e42-48a6-802a-553364d32a9b-host\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.614995 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b84d71a3-4e42-48a6-802a-553364d32a9b-serviceca\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.615043 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86mzf\" (UniqueName: \"kubernetes.io/projected/b84d71a3-4e42-48a6-802a-553364d32a9b-kube-api-access-86mzf\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.615086 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b84d71a3-4e42-48a6-802a-553364d32a9b-host\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.617090 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b84d71a3-4e42-48a6-802a-553364d32a9b-serviceca\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.621000 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.636662 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.641584 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86mzf\" (UniqueName: \"kubernetes.io/projected/b84d71a3-4e42-48a6-802a-553364d32a9b-kube-api-access-86mzf\") pod \"node-ca-8plsb\" (UID: \"b84d71a3-4e42-48a6-802a-553364d32a9b\") " pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.654656 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.674112 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.705801 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.705866 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.705888 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.705916 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.705941 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.746973 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8plsb" Dec 03 06:50:35 crc kubenswrapper[4946]: W1203 06:50:35.763614 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb84d71a3_4e42_48a6_802a_553364d32a9b.slice/crio-cd0cb65232571978ca5945310de691264c2a00d8f6aca59c05b70d19364994e6 WatchSource:0}: Error finding container cd0cb65232571978ca5945310de691264c2a00d8f6aca59c05b70d19364994e6: Status 404 returned error can't find the container with id cd0cb65232571978ca5945310de691264c2a00d8f6aca59c05b70d19364994e6 Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.808805 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.808864 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.808882 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.808909 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.808929 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.812222 4946 generic.go:334] "Generic (PLEG): container finished" podID="81237850-a445-4887-86e0-23bb0fa052c2" containerID="73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452" exitCode=0 Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.812390 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerDied","Data":"73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.814604 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8plsb" event={"ID":"b84d71a3-4e42-48a6-802a-553364d32a9b","Type":"ContainerStarted","Data":"cd0cb65232571978ca5945310de691264c2a00d8f6aca59c05b70d19364994e6"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.822460 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.833394 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.855058 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.873415 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.891266 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.906635 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.912255 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.912297 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.912309 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.912327 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.912343 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:35Z","lastTransitionTime":"2025-12-03T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.929392 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.942298 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.955556 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:35 crc kubenswrapper[4946]: I1203 06:50:35.995207 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.014503 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.014558 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.014581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.014608 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.014628 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.043243 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.074916 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.122280 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.123153 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.123212 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.123224 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.123359 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.123432 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.156714 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.193517 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.225785 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.225839 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.225855 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.225879 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.225895 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.235806 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.329869 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.329958 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.329975 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.330000 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.330015 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.433804 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.433867 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.433881 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.433906 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.433926 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.538028 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.538130 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.538160 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.538198 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.538238 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.591773 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:36 crc kubenswrapper[4946]: E1203 06:50:36.591994 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.592171 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:36 crc kubenswrapper[4946]: E1203 06:50:36.592491 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.640763 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.640835 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.640849 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.640872 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.640887 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.744478 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.744534 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.744552 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.744578 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.744597 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.829081 4946 generic.go:334] "Generic (PLEG): container finished" podID="81237850-a445-4887-86e0-23bb0fa052c2" containerID="c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd" exitCode=0 Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.829125 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerDied","Data":"c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.831393 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8plsb" event={"ID":"b84d71a3-4e42-48a6-802a-553364d32a9b","Type":"ContainerStarted","Data":"dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.847708 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.847809 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.847828 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.847854 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.847897 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.854250 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.873773 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.904607 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.921613 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.934205 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.951171 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.951208 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.951220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.951239 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.951260 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:36Z","lastTransitionTime":"2025-12-03T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.965168 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:36 crc kubenswrapper[4946]: I1203 06:50:36.982433 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:36Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.003078 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.023211 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.046384 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.053973 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.054024 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.054038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.054061 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.054074 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.065139 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.083160 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.105351 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.124896 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.150503 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.156999 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.157027 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.157037 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.157053 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.157065 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.182931 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.200653 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.214070 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.231408 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.254408 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.259725 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.259792 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.259819 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.259840 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.259855 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.272275 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.286657 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.310946 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.332670 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.362488 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.362533 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.362544 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.362566 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.362579 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.370914 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.392131 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.404159 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.422713 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.425277 4946 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 03 06:50:37 crc kubenswrapper[4946]: E1203 06:50:37.425631 4946 request.go:1255] Unexpected error when reading response body: read tcp 38.102.83.174:46532->38.102.83.174:6443: use of closed network connection Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.425874 4946 status_manager.go:851] "Failed to get status for pod" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="unexpected error when reading response body. Please retry. Original error: read tcp 38.102.83.174:46532->38.102.83.174:6443: use of closed network connection" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.451826 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.464780 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.465018 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.465100 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.465172 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.465233 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.569099 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.569180 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.569200 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.569228 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.569249 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.592853 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:37 crc kubenswrapper[4946]: E1203 06:50:37.593117 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.609503 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.631816 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.649455 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.665506 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.674341 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.674379 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.674390 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.674408 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.674420 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.686488 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.706939 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.723120 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.738500 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.777125 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.777165 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.777175 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.777192 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.777204 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.789890 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.819326 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.839565 4946 generic.go:334] "Generic (PLEG): container finished" podID="81237850-a445-4887-86e0-23bb0fa052c2" containerID="2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6" exitCode=0 Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.839664 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerDied","Data":"2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.846733 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.854926 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.881824 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.882083 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.882105 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.882134 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.882154 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.898205 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.939007 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.972932 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.985463 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.985508 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.985526 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.985549 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:37 crc kubenswrapper[4946]: I1203 06:50:37.985567 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:37Z","lastTransitionTime":"2025-12-03T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.019010 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.053312 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.088516 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.088584 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.088605 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.088631 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.088649 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.095648 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.130920 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.181085 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.191509 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.191567 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.191580 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.191596 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.191610 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.212562 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.253214 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.295264 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.295358 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.295389 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.295418 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.295438 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.298945 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.338033 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.378300 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.398389 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.398464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.398475 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.398492 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.398505 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.414434 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.470098 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.496808 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.501716 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.501822 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.501850 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.501882 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.501907 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.535776 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.578141 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.592254 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.592324 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:38 crc kubenswrapper[4946]: E1203 06:50:38.592503 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:38 crc kubenswrapper[4946]: E1203 06:50:38.592636 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.604921 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.604992 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.605007 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.605038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.605052 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.621438 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.707903 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.707967 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.707985 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.708012 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.708032 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.810852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.810899 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.810911 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.810932 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.810944 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.858869 4946 generic.go:334] "Generic (PLEG): container finished" podID="81237850-a445-4887-86e0-23bb0fa052c2" containerID="62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236" exitCode=0 Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.858987 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerDied","Data":"62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.883502 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.900397 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.914962 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.917015 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.917071 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.917086 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.917111 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.917207 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:38Z","lastTransitionTime":"2025-12-03T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.929864 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.952490 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.970132 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:38 crc kubenswrapper[4946]: I1203 06:50:38.980799 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:38Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.002378 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.016795 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.020935 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.020986 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.020999 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.021023 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.021038 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.033281 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.052445 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.057999 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.058049 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.058061 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.058081 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.058096 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: E1203 06:50:39.074011 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.077774 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.077810 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.077821 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.077839 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.077852 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: E1203 06:50:39.092428 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.092462 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.097415 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.097458 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.097481 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.097498 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.097509 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: E1203 06:50:39.110817 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.115763 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.115796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.115817 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.115835 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.115847 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: E1203 06:50:39.133711 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.135770 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.138871 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.138920 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.138934 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.138957 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.138969 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: E1203 06:50:39.151262 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: E1203 06:50:39.151378 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.153276 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.153336 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.153356 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.153380 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.153394 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.175793 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.213117 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.255820 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.255857 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.255867 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.255886 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.255900 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.358401 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.358449 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.358463 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.358483 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.358498 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.461688 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.461813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.461836 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.461866 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.461885 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.569718 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.570322 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.570347 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.570374 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.570393 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.592293 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:39 crc kubenswrapper[4946]: E1203 06:50:39.592477 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.674285 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.674342 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.674359 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.674379 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.674392 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.778232 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.778307 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.778324 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.778353 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.778374 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.869611 4946 generic.go:334] "Generic (PLEG): container finished" podID="81237850-a445-4887-86e0-23bb0fa052c2" containerID="f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86" exitCode=0 Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.869701 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerDied","Data":"f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.877409 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.878108 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.878154 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.881438 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.881492 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.881511 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.881538 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.881571 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.886984 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.914983 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.915978 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.929095 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.936503 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.952158 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.965853 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.979212 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.984735 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.984810 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.984822 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.984844 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.984861 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:39Z","lastTransitionTime":"2025-12-03T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:39 crc kubenswrapper[4946]: I1203 06:50:39.992135 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:39Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.007279 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.026271 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.039154 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.052528 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.067314 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.081558 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.088092 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.088124 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.088133 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.088147 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.088157 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.094516 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.111358 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.130878 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.141830 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.152330 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.167089 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.179569 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.190958 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.191018 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.191031 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.191078 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.191094 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.199042 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.216306 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.234626 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.246256 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.260214 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.277469 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.293395 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.293433 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.293445 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.293465 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.293478 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.294074 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.331217 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.371424 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.395352 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.395391 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.395399 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.395417 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.395428 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.414657 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.498708 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.498922 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.499075 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.499217 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.499332 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.592517 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.592517 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:40 crc kubenswrapper[4946]: E1203 06:50:40.592713 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:40 crc kubenswrapper[4946]: E1203 06:50:40.592912 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.602553 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.602620 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.602639 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.602666 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.602685 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.707178 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.707565 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.707587 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.707610 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.707627 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.811131 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.811199 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.811216 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.811241 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.811259 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.886506 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" event={"ID":"81237850-a445-4887-86e0-23bb0fa052c2","Type":"ContainerStarted","Data":"047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.886574 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.907245 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.914698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.914800 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.914825 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.914859 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.914883 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:40Z","lastTransitionTime":"2025-12-03T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.929166 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.943280 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.967335 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.984097 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:40 crc kubenswrapper[4946]: I1203 06:50:40.998604 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.018183 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.018228 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.018241 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.018258 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.018270 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.020140 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.035950 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.052905 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.068713 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.088383 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.103817 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.121180 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.121249 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.121260 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.121278 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.121306 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.122691 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.137549 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.153479 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.224824 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.224949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.224968 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.224994 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.225013 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.329678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.329778 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.329797 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.329821 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.329838 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.432842 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.432910 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.432934 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.432964 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.432986 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.536392 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.536455 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.536478 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.536509 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.536532 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.593546 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:41 crc kubenswrapper[4946]: E1203 06:50:41.593734 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.639134 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.639184 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.639201 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.639225 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.639246 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.741988 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.742028 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.742037 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.742052 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.742061 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.844868 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.844906 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.844918 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.844936 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.844947 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.889896 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.948013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.948056 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.948070 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.948088 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:41 crc kubenswrapper[4946]: I1203 06:50:41.948098 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:41Z","lastTransitionTime":"2025-12-03T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.051153 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.051203 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.051218 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.051238 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.051249 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.154580 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.154646 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.154665 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.154688 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.154707 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.257439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.257513 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.257535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.257564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.257587 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.311445 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.311836 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:50:58.311804975 +0000 UTC m=+51.108495124 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.360126 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.360174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.360189 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.360206 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.360218 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.413029 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.413094 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.413121 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.413154 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413237 4946 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413264 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413287 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413299 4946 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413248 4946 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413331 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:58.413311643 +0000 UTC m=+51.210001772 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413350 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:58.413340684 +0000 UTC m=+51.210030793 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413363 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:58.413356584 +0000 UTC m=+51.210046693 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413361 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413408 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413424 4946 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.413509 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:58.413481098 +0000 UTC m=+51.210171257 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.463930 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.463990 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.464002 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.464021 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.464034 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.567570 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.567633 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.567649 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.567672 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.567689 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.592056 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.592173 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.592319 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:42 crc kubenswrapper[4946]: E1203 06:50:42.592412 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.670143 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.670187 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.670198 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.670214 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.670228 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.773833 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.773920 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.773941 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.773967 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.773986 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.877170 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.877241 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.877264 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.877294 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.877316 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.896646 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/0.log" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.900863 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff" exitCode=1 Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.900930 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.902185 4946 scope.go:117] "RemoveContainer" containerID="b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.922514 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.949414 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.964800 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.981604 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.981673 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.981697 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.981731 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.981798 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:42Z","lastTransitionTime":"2025-12-03T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:42 crc kubenswrapper[4946]: I1203 06:50:42.984915 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.007215 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.027407 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.042689 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.083249 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.084812 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.084869 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.084884 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.084906 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.084924 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.116211 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.137035 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.149108 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.165059 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.177396 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.187699 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.187779 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.187790 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.187809 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.187819 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.192069 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.206581 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.290106 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.290176 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.290187 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.290204 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.290216 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.392973 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.393039 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.393049 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.393074 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.393088 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.496599 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.496692 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.496719 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.496796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.496820 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.592032 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:43 crc kubenswrapper[4946]: E1203 06:50:43.593031 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.598947 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.599008 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.599018 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.599031 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.599041 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.702173 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.702235 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.702250 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.702271 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.702289 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.805553 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.805586 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.805595 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.805608 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.805617 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.907623 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/0.log" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.909099 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.909338 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.909361 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.909389 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.909411 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:43Z","lastTransitionTime":"2025-12-03T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.911724 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d"} Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.911977 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.933497 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.951577 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.969451 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:43 crc kubenswrapper[4946]: I1203 06:50:43.982176 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.011313 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.013222 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.013269 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.013286 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.013309 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.013329 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.031710 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.038582 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.047966 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.067490 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.087361 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.107674 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.116326 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.116369 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.116378 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.116395 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.116406 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.124205 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.139108 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.158182 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.174553 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.194142 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.212191 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.219936 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.220001 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.220285 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.220362 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.220383 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.228780 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.244236 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.255227 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.267727 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.285531 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.296650 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.310844 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.321788 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.323967 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.324033 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.324053 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.324078 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.324093 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.344173 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.357271 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.385179 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.398900 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.414417 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.426855 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.426882 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.426891 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.426906 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.426916 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.431373 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.528918 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.528966 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.528978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.528993 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.529003 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.591849 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.591892 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:44 crc kubenswrapper[4946]: E1203 06:50:44.592072 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:44 crc kubenswrapper[4946]: E1203 06:50:44.592204 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.632698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.633022 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.633158 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.633282 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.633415 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.735896 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.735938 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.735950 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.735967 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.735979 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.838993 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.839061 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.839083 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.839114 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.839137 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.917651 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/1.log" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.918588 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/0.log" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.922492 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d" exitCode=1 Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.922538 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.922596 4946 scope.go:117] "RemoveContainer" containerID="b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.924089 4946 scope.go:117] "RemoveContainer" containerID="31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d" Dec 03 06:50:44 crc kubenswrapper[4946]: E1203 06:50:44.924509 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.940623 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.941636 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.941673 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.941683 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.941702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.941714 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:44Z","lastTransitionTime":"2025-12-03T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.955849 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.967603 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:44 crc kubenswrapper[4946]: I1203 06:50:44.991042 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.004931 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.028329 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.042518 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.045027 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.045117 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.045150 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.045174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.045190 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.059551 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.073945 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.093010 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.109679 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.125752 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.140889 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.147728 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.147781 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.147792 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.147811 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.147824 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.155092 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.173132 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.250943 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.251026 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.251048 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.251079 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.251097 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.354768 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.354837 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.354856 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.354889 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.354906 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.458058 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.458108 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.458123 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.458144 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.458159 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.561135 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.561205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.561223 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.561250 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.561269 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.592079 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:45 crc kubenswrapper[4946]: E1203 06:50:45.592270 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.664759 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.664828 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.664841 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.664859 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.664871 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.768118 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.768157 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.768167 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.768181 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.768190 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.870387 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.870450 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.870465 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.870488 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.870501 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.928352 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/1.log" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.973987 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.974047 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.974066 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.974093 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:45 crc kubenswrapper[4946]: I1203 06:50:45.974113 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:45Z","lastTransitionTime":"2025-12-03T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.062351 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh"] Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.063216 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.066464 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.066498 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.078983 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.079037 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.079055 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.079077 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.079097 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.086456 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.103701 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.137819 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.153884 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a3371460-14a2-409b-9b8e-603481a252ab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.153970 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q58cn\" (UniqueName: \"kubernetes.io/projected/a3371460-14a2-409b-9b8e-603481a252ab-kube-api-access-q58cn\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.154071 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a3371460-14a2-409b-9b8e-603481a252ab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.154155 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a3371460-14a2-409b-9b8e-603481a252ab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.158384 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.174255 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.182369 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.182465 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.182484 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.182508 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.182525 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.208604 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.229853 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.250543 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.254995 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a3371460-14a2-409b-9b8e-603481a252ab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.255103 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a3371460-14a2-409b-9b8e-603481a252ab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.255203 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a3371460-14a2-409b-9b8e-603481a252ab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.255240 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q58cn\" (UniqueName: \"kubernetes.io/projected/a3371460-14a2-409b-9b8e-603481a252ab-kube-api-access-q58cn\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.256170 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a3371460-14a2-409b-9b8e-603481a252ab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.256552 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a3371460-14a2-409b-9b8e-603481a252ab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.263910 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a3371460-14a2-409b-9b8e-603481a252ab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.274151 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.286622 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.286665 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q58cn\" (UniqueName: \"kubernetes.io/projected/a3371460-14a2-409b-9b8e-603481a252ab-kube-api-access-q58cn\") pod \"ovnkube-control-plane-749d76644c-pxmvh\" (UID: \"a3371460-14a2-409b-9b8e-603481a252ab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.286699 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.286725 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.286791 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.286817 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.302848 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.322060 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.340649 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.362591 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.381271 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.384441 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.390110 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.390174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.390198 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.390232 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.390255 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.400851 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: W1203 06:50:46.406018 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3371460_14a2_409b_9b8e_603481a252ab.slice/crio-cfc68c527a98ca7993e63ea33a6c2e6bc0ab8e6acf3b5c6c416f70efdf0de76b WatchSource:0}: Error finding container cfc68c527a98ca7993e63ea33a6c2e6bc0ab8e6acf3b5c6c416f70efdf0de76b: Status 404 returned error can't find the container with id cfc68c527a98ca7993e63ea33a6c2e6bc0ab8e6acf3b5c6c416f70efdf0de76b Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.426500 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.493349 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.493411 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.493434 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.493464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.493489 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.592336 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:46 crc kubenswrapper[4946]: E1203 06:50:46.592525 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.592347 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:46 crc kubenswrapper[4946]: E1203 06:50:46.593070 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.597608 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.597802 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.597831 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.597916 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.597943 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.702271 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.702318 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.702331 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.702351 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.702367 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.805293 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.805359 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.805378 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.805404 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.805425 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.908859 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.908910 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.908920 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.908940 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.908952 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:46Z","lastTransitionTime":"2025-12-03T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:46 crc kubenswrapper[4946]: I1203 06:50:46.938141 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" event={"ID":"a3371460-14a2-409b-9b8e-603481a252ab","Type":"ContainerStarted","Data":"cfc68c527a98ca7993e63ea33a6c2e6bc0ab8e6acf3b5c6c416f70efdf0de76b"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.012446 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.012511 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.012525 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.012548 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.012561 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.115723 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.115827 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.115850 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.115919 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.115949 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.215111 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hkpdf"] Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.215710 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:47 crc kubenswrapper[4946]: E1203 06:50:47.215821 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.218113 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.218162 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.218175 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.218196 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.218229 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.230759 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.243382 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.258810 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.275559 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.293294 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.307666 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.320929 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.320983 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.320996 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.321015 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.321028 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.325994 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.338569 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.351435 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.367170 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.367232 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p5zt\" (UniqueName: \"kubernetes.io/projected/2e5abff8-4252-4371-803d-d241c81c5910-kube-api-access-9p5zt\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.378575 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.392790 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.406294 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.423631 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.423690 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.423709 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.423786 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.423811 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.427375 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.445509 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.458822 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.468040 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p5zt\" (UniqueName: \"kubernetes.io/projected/2e5abff8-4252-4371-803d-d241c81c5910-kube-api-access-9p5zt\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.468114 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:47 crc kubenswrapper[4946]: E1203 06:50:47.468234 4946 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:47 crc kubenswrapper[4946]: E1203 06:50:47.468585 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs podName:2e5abff8-4252-4371-803d-d241c81c5910 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:47.968273322 +0000 UTC m=+40.764963431 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs") pod "network-metrics-daemon-hkpdf" (UID: "2e5abff8-4252-4371-803d-d241c81c5910") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.472101 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.487343 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p5zt\" (UniqueName: \"kubernetes.io/projected/2e5abff8-4252-4371-803d-d241c81c5910-kube-api-access-9p5zt\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.491354 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.527172 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.527237 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.527253 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.527272 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.527283 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.591988 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:47 crc kubenswrapper[4946]: E1203 06:50:47.592139 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.627192 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.629608 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.629662 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.629679 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.629706 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.629723 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.650397 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.676445 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.696986 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.717173 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.732938 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.732978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.732987 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.733005 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.733016 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.733090 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.744001 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.757854 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.768907 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.787101 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.811704 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.825269 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.835145 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.835188 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.835201 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.835221 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.835234 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.847230 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.858159 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.880631 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.897167 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.907092 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.937867 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.937918 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.937929 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.937947 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.937960 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:47Z","lastTransitionTime":"2025-12-03T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.943477 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" event={"ID":"a3371460-14a2-409b-9b8e-603481a252ab","Type":"ContainerStarted","Data":"dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.943547 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" event={"ID":"a3371460-14a2-409b-9b8e-603481a252ab","Type":"ContainerStarted","Data":"960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9"} Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.960525 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.974172 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:47 crc kubenswrapper[4946]: E1203 06:50:47.974307 4946 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:47 crc kubenswrapper[4946]: E1203 06:50:47.974360 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs podName:2e5abff8-4252-4371-803d-d241c81c5910 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:48.974346583 +0000 UTC m=+41.771036692 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs") pod "network-metrics-daemon-hkpdf" (UID: "2e5abff8-4252-4371-803d-d241c81c5910") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:47 crc kubenswrapper[4946]: I1203 06:50:47.985931 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.001869 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.020183 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.037572 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.040283 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.040340 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.040356 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.040390 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.040407 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.075290 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.106386 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.119638 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.130677 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.143093 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.143130 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.143140 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.143159 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.143171 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.148616 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.162375 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.172972 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.186402 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.195967 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.212215 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.221929 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.240319 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.245018 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.245048 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.245057 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.245071 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.245081 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.347704 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.347798 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.347809 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.347823 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.347832 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.450395 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.450437 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.450447 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.450463 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.450476 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.553824 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.553896 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.553914 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.553942 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.553960 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.591800 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.591833 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.591902 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:48 crc kubenswrapper[4946]: E1203 06:50:48.592006 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:50:48 crc kubenswrapper[4946]: E1203 06:50:48.592145 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:48 crc kubenswrapper[4946]: E1203 06:50:48.592269 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.657825 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.657904 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.657931 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.657960 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.657979 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.761605 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.761670 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.761691 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.761715 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.761733 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.865371 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.865448 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.865467 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.865498 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.865519 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.968935 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.969010 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.969021 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.969039 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.969051 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:48Z","lastTransitionTime":"2025-12-03T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:48 crc kubenswrapper[4946]: I1203 06:50:48.987832 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:48 crc kubenswrapper[4946]: E1203 06:50:48.988101 4946 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:48 crc kubenswrapper[4946]: E1203 06:50:48.988203 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs podName:2e5abff8-4252-4371-803d-d241c81c5910 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:50.988178218 +0000 UTC m=+43.784868367 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs") pod "network-metrics-daemon-hkpdf" (UID: "2e5abff8-4252-4371-803d-d241c81c5910") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.072018 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.072058 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.072067 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.072083 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.072092 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.175293 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.175350 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.175359 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.175377 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.175387 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.278471 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.278529 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.278540 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.278558 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.278570 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.381355 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.381413 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.381422 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.381456 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.381468 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.433091 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.433140 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.433159 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.433184 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.433203 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: E1203 06:50:49.448945 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.455050 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.455122 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.455144 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.455174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.455195 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: E1203 06:50:49.476077 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.481435 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.481519 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.481553 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.481619 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.481644 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: E1203 06:50:49.503108 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.508202 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.508251 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.508267 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.508291 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.508309 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: E1203 06:50:49.527843 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.533228 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.533287 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.533327 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.533363 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.533388 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: E1203 06:50:49.553858 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:49 crc kubenswrapper[4946]: E1203 06:50:49.554085 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.556313 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.556413 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.556480 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.556511 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.556529 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.592661 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:49 crc kubenswrapper[4946]: E1203 06:50:49.592918 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.659163 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.659220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.659238 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.659261 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.659280 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.762640 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.762690 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.762701 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.762720 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.762732 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.865980 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.866035 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.866050 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.866071 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.866085 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.969108 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.969163 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.969183 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.969204 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:49 crc kubenswrapper[4946]: I1203 06:50:49.969220 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:49Z","lastTransitionTime":"2025-12-03T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.071776 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.071818 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.071829 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.071847 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.071860 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.174910 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.174981 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.175000 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.175026 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.175044 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.277683 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.277803 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.277829 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.277861 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.277884 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.381135 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.381201 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.381219 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.381247 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.381265 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.485350 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.485412 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.485430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.485454 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.485472 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.588194 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.588254 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.588277 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.588303 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.588321 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.591720 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.591812 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:50 crc kubenswrapper[4946]: E1203 06:50:50.591889 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:50 crc kubenswrapper[4946]: E1203 06:50:50.591985 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.591826 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:50 crc kubenswrapper[4946]: E1203 06:50:50.592200 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.691840 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.691896 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.691915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.691941 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.691960 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.795046 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.795110 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.795130 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.795157 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.795176 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.898968 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.899047 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.899069 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.899095 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:50 crc kubenswrapper[4946]: I1203 06:50:50.899114 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:50Z","lastTransitionTime":"2025-12-03T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.002442 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.002511 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.002532 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.002561 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.002582 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.008703 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:51 crc kubenswrapper[4946]: E1203 06:50:51.008943 4946 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:51 crc kubenswrapper[4946]: E1203 06:50:51.009030 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs podName:2e5abff8-4252-4371-803d-d241c81c5910 nodeName:}" failed. No retries permitted until 2025-12-03 06:50:55.009008241 +0000 UTC m=+47.805698380 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs") pod "network-metrics-daemon-hkpdf" (UID: "2e5abff8-4252-4371-803d-d241c81c5910") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.105543 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.105622 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.105634 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.105654 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.105690 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.209236 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.209312 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.209335 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.209365 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.209388 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.312061 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.312223 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.312253 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.312286 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.312310 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.415673 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.415729 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.415761 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.415782 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.415796 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.518666 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.518712 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.518722 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.518767 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.518783 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.592029 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:51 crc kubenswrapper[4946]: E1203 06:50:51.592188 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.621400 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.621468 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.621484 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.621512 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.621533 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.725051 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.725104 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.725122 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.725147 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.725166 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.827840 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.827898 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.827915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.827943 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.827966 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.931494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.931551 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.931564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.931585 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:51 crc kubenswrapper[4946]: I1203 06:50:51.931599 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:51Z","lastTransitionTime":"2025-12-03T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.034485 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.034583 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.034612 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.034647 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.034675 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.137991 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.138038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.138048 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.138065 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.138076 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.241147 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.241194 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.241205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.241221 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.241232 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.344235 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.344295 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.344311 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.344335 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.344370 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.447635 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.447677 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.447685 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.447704 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.447715 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.550597 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.550642 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.550653 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.550669 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.550680 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.592205 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.592301 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.592385 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:52 crc kubenswrapper[4946]: E1203 06:50:52.592509 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:52 crc kubenswrapper[4946]: E1203 06:50:52.592682 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:50:52 crc kubenswrapper[4946]: E1203 06:50:52.592776 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.654182 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.654252 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.654273 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.654336 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.654355 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.757784 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.757850 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.757868 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.757891 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.757908 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.861411 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.861465 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.861499 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.861519 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.861531 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.964432 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.964501 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.964513 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.964532 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:52 crc kubenswrapper[4946]: I1203 06:50:52.964543 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:52Z","lastTransitionTime":"2025-12-03T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.066912 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.066953 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.066963 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.066978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.066989 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.169803 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.169905 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.169924 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.169984 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.170004 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.273134 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.273196 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.273218 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.273244 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.273262 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.377189 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.377244 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.377257 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.377277 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.377296 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.480566 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.480625 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.480640 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.480660 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.480674 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.584411 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.584479 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.584499 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.584526 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.584544 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.591828 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:53 crc kubenswrapper[4946]: E1203 06:50:53.592039 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.687389 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.687449 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.687466 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.687490 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.687507 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.791394 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.791439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.791448 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.791466 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.791477 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.894621 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.894793 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.894813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.894837 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.894855 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.997732 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.997840 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.997858 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.998008 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:53 crc kubenswrapper[4946]: I1203 06:50:53.998032 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:53Z","lastTransitionTime":"2025-12-03T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.100431 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.100510 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.100532 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.100558 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.100579 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.203585 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.203628 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.203640 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.203657 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.203674 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.307036 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.307110 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.307135 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.307166 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.307189 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.410680 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.410782 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.410810 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.410839 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.410860 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.513796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.513867 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.513887 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.513915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.513934 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.592651 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.592700 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.592811 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:54 crc kubenswrapper[4946]: E1203 06:50:54.592808 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:54 crc kubenswrapper[4946]: E1203 06:50:54.592945 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:50:54 crc kubenswrapper[4946]: E1203 06:50:54.592987 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.617414 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.617494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.617514 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.617536 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.617586 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.721317 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.721374 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.721397 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.721423 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.721446 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.824691 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.825285 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.825381 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.825488 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.825573 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.930097 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.930163 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.930185 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.930216 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:54 crc kubenswrapper[4946]: I1203 06:50:54.930244 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:54Z","lastTransitionTime":"2025-12-03T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.033045 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.033124 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.033152 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.033184 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.033210 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.049941 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:55 crc kubenswrapper[4946]: E1203 06:50:55.050161 4946 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:55 crc kubenswrapper[4946]: E1203 06:50:55.050239 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs podName:2e5abff8-4252-4371-803d-d241c81c5910 nodeName:}" failed. No retries permitted until 2025-12-03 06:51:03.050219287 +0000 UTC m=+55.846909396 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs") pod "network-metrics-daemon-hkpdf" (UID: "2e5abff8-4252-4371-803d-d241c81c5910") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.136887 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.136952 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.136968 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.136988 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.137001 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.239800 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.240316 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.240385 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.240460 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.240541 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.343678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.343829 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.343852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.343884 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.344047 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.446997 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.447051 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.447066 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.447087 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.447102 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.550529 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.550563 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.550572 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.550585 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.550595 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.592687 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:55 crc kubenswrapper[4946]: E1203 06:50:55.592976 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.653724 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.653869 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.653898 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.653929 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.653954 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.757217 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.757290 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.757313 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.757346 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.757373 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.859939 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.859999 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.860018 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.860041 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.860063 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.903525 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.918646 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.929498 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.945534 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.963796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.963860 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.963953 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.963985 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.964008 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:55Z","lastTransitionTime":"2025-12-03T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:55 crc kubenswrapper[4946]: I1203 06:50:55.983878 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.002228 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.018458 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.041919 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.061184 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.066862 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.066896 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.066907 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.066925 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.066936 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.081287 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.100765 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.120867 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.134570 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.150091 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.164172 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.169707 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.169774 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.169787 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.169802 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.169813 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.178532 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.189792 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.202609 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.214185 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.272527 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.272554 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.272565 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.272581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.272592 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.375789 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.375842 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.375858 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.375880 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.375896 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.478505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.478556 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.478565 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.478577 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.478586 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.581484 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.581533 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.581550 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.581573 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.581590 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.591708 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.591712 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:56 crc kubenswrapper[4946]: E1203 06:50:56.591903 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.591966 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:56 crc kubenswrapper[4946]: E1203 06:50:56.592128 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:56 crc kubenswrapper[4946]: E1203 06:50:56.592274 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.684414 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.684472 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.684485 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.684500 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.684512 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.788054 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.788120 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.788143 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.788176 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.788199 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.890420 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.890467 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.890477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.890490 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.890499 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.992564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.992860 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.992969 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.993051 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:56 crc kubenswrapper[4946]: I1203 06:50:56.993118 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:56Z","lastTransitionTime":"2025-12-03T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.096866 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.097206 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.097291 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.097377 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.097450 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.201079 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.201136 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.201146 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.201161 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.201170 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.307293 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.307349 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.307367 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.307421 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.307439 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.410356 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.410436 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.410460 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.410490 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.410516 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.513907 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.513953 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.513966 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.513987 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.514000 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.592517 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:57 crc kubenswrapper[4946]: E1203 06:50:57.592806 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.614856 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.616312 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.616422 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.616442 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.616466 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.616484 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.628271 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.659271 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1f26ea8458bc4ba1b562348bfb3b8189d0fe15496bbdf0fd6f9db9e725843ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:42Z\\\",\\\"message\\\":\\\"06:50:42.408478 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:50:42.408495 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:42.408529 6225 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:42.408544 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:50:42.408558 6225 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:50:42.408555 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:42.408565 6225 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:42.408568 6225 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:42.408578 6225 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:50:42.408583 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:42.408588 6225 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:42.408609 6225 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:42.408617 6225 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:42.408700 6225 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:42.408895 6225 factory.go:656] Stopping watch factory\\\\nI1203 06:50:42.408975 6225 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.680226 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.697868 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.720023 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.720057 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.720065 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.720077 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.720086 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.722419 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.738314 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.758238 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.772947 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.789277 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.810877 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.823002 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.823064 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.823082 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.823107 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.823126 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.829221 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.852795 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.869894 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.887475 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.905826 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.926611 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.926649 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.926662 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.926679 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.926693 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:57Z","lastTransitionTime":"2025-12-03T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.931330 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:57 crc kubenswrapper[4946]: I1203 06:50:57.947371 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.029678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.029838 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.029864 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.029894 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.029970 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.132750 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.132799 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.132808 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.132826 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.132838 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.235768 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.235809 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.235822 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.235838 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.235848 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.338654 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.338710 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.338728 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.338783 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.338848 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.382136 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.382304 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:51:30.382261103 +0000 UTC m=+83.178951252 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.441361 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.441414 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.441432 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.441456 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.441473 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.483350 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.483471 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.483502 4946 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.483556 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.483613 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:51:30.483570576 +0000 UTC m=+83.280260725 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.483659 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.483858 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.483892 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.483920 4946 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.484004 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:51:30.483981646 +0000 UTC m=+83.280671785 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.484158 4946 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.484356 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:51:30.484278134 +0000 UTC m=+83.280968293 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.484544 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.484602 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.484619 4946 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.484702 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:51:30.484684995 +0000 UTC m=+83.281375144 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.544524 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.545097 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.545300 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.545498 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.545652 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.592592 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.592637 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.592686 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.593529 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.593819 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.593869 4946 scope.go:117] "RemoveContainer" containerID="31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d" Dec 03 06:50:58 crc kubenswrapper[4946]: E1203 06:50:58.593915 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.615272 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.632732 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.649822 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.649877 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.649893 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.649920 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.649938 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.669369 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.689276 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.705910 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.742537 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.753083 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.753151 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.753170 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.753198 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.753217 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.760144 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.780287 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.802693 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.826154 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.844656 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.855625 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.855660 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.855668 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.855682 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.855691 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.858033 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.872921 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.883839 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.896676 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.908131 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.920455 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.931183 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.958375 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.958411 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.958419 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.958432 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.958441 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:58Z","lastTransitionTime":"2025-12-03T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.984571 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/1.log" Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.987898 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e"} Dec 03 06:50:58 crc kubenswrapper[4946]: I1203 06:50:58.988060 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.008298 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.023335 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.040015 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.060983 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.061965 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.062006 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.062017 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.062034 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.062050 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.080475 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.095945 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.106847 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.119994 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.130043 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.141175 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.150529 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.165086 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.165141 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.165154 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.165182 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.165193 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.180037 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.192809 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.203085 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.221604 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.234790 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.247284 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.259562 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.267296 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.267335 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.267346 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.267363 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.267385 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.370597 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.370630 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.370638 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.370651 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.370659 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.474362 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.474433 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.474458 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.474489 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.474514 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.577854 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.577924 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.577946 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.578013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.578039 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.591876 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:50:59 crc kubenswrapper[4946]: E1203 06:50:59.592042 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.636612 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.636660 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.636672 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.636687 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.636698 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: E1203 06:50:59.650900 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.654933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.654981 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.654996 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.655019 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.655034 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: E1203 06:50:59.674543 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.679110 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.679175 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.679195 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.679217 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.679232 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: E1203 06:50:59.697709 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.702421 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.702462 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.702474 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.702493 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.702507 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: E1203 06:50:59.719882 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.724309 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.724523 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.724622 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.724705 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.724796 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: E1203 06:50:59.742219 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:50:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:50:59 crc kubenswrapper[4946]: E1203 06:50:59.742406 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.744183 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.744221 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.744236 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.744261 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.744274 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.847170 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.847232 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.847250 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.847274 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.847294 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.950266 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.950320 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.950343 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.950372 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.950397 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:50:59Z","lastTransitionTime":"2025-12-03T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.994266 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/2.log" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.995285 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/1.log" Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.999080 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e" exitCode=1 Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.999138 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e"} Dec 03 06:50:59 crc kubenswrapper[4946]: I1203 06:50:59.999198 4946 scope.go:117] "RemoveContainer" containerID="31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.000923 4946 scope.go:117] "RemoveContainer" containerID="ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e" Dec 03 06:51:00 crc kubenswrapper[4946]: E1203 06:51:00.001335 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.027149 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.048510 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.052808 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.052888 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.052915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.052947 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.052970 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.068896 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.102508 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.115988 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.137363 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.153980 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.156124 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.156182 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.156201 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.156227 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.156245 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.171381 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.189256 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.205779 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.225222 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.249618 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.259092 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.259128 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.259138 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.259155 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.259166 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.261774 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.273059 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.301337 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31a315120aaf92a0149e62845607428ca5d732955831d5588d576467d678e74d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"message\\\":\\\"ator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nI1203 06:50:43.861207 6357 services_controller.go:443] Built service openshift-multus/multus-admission-controller LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861213 6357 services_controller.go:443] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.168\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 06:50:43.861238 6357 services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.313140 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.326178 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.340461 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.361347 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.361390 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.361419 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.361436 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.361444 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.464060 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.464120 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.464138 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.464161 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.464180 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.567122 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.567176 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.567188 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.567207 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.567219 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.592409 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.592482 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:00 crc kubenswrapper[4946]: E1203 06:51:00.592567 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.592429 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:00 crc kubenswrapper[4946]: E1203 06:51:00.592704 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:00 crc kubenswrapper[4946]: E1203 06:51:00.592786 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.669593 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.669640 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.669648 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.669663 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.669674 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.772227 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.772284 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.772304 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.772327 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.772344 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.875274 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.875322 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.875336 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.875354 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.875367 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.978309 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.978371 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.978398 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.978430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:00 crc kubenswrapper[4946]: I1203 06:51:00.978453 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:00Z","lastTransitionTime":"2025-12-03T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.008100 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/2.log" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.081506 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.081936 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.082134 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.082314 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.082511 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.186174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.186220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.186232 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.186248 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.186260 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.289418 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.289478 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.289494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.289519 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.289535 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.392405 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.392457 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.392474 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.392497 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.392515 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.495420 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.495483 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.495505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.495535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.495560 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.592073 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:01 crc kubenswrapper[4946]: E1203 06:51:01.592242 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.599015 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.599066 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.599100 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.599119 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.599151 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.703236 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.703320 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.703332 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.703352 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.703367 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.807279 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.807344 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.807367 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.807397 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.807421 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.910262 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.910320 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.910337 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.910359 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:01 crc kubenswrapper[4946]: I1203 06:51:01.910376 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:01Z","lastTransitionTime":"2025-12-03T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.012919 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.013137 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.013162 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.013353 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.013432 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.115651 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.115886 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.115958 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.116021 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.116081 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.218915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.219253 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.219452 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.219593 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.219723 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.322425 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.322879 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.323017 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.323169 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.323302 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.426381 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.426433 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.426443 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.426459 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.426469 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.528911 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.528986 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.529009 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.529035 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.529054 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.592688 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:02 crc kubenswrapper[4946]: E1203 06:51:02.592869 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.593001 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.592688 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:02 crc kubenswrapper[4946]: E1203 06:51:02.593221 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:02 crc kubenswrapper[4946]: E1203 06:51:02.593350 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.632023 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.632068 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.632079 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.632093 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.632102 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.734604 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.734634 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.734663 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.734680 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.734692 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.791680 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.792816 4946 scope.go:117] "RemoveContainer" containerID="ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e" Dec 03 06:51:02 crc kubenswrapper[4946]: E1203 06:51:02.793084 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.809364 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.825525 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.837468 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.837523 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.837541 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.837564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.837583 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.839636 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.852567 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.865250 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.895883 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.912405 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.924163 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.941141 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.941184 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.941197 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.941219 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.941233 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:02Z","lastTransitionTime":"2025-12-03T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.941812 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:02 crc kubenswrapper[4946]: I1203 06:51:02.986344 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.006868 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.021528 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.040583 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.043070 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.043120 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.043134 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.043154 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.043166 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.052535 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.070020 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.094500 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.110151 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.126591 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.130927 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:03 crc kubenswrapper[4946]: E1203 06:51:03.131095 4946 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:51:03 crc kubenswrapper[4946]: E1203 06:51:03.131151 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs podName:2e5abff8-4252-4371-803d-d241c81c5910 nodeName:}" failed. No retries permitted until 2025-12-03 06:51:19.131135607 +0000 UTC m=+71.927825706 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs") pod "network-metrics-daemon-hkpdf" (UID: "2e5abff8-4252-4371-803d-d241c81c5910") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.146338 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.146385 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.146398 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.146417 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.146432 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.249348 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.249408 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.249426 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.249452 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.249469 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.352634 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.352674 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.352685 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.352703 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.352717 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.455995 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.456072 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.456099 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.456131 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.456153 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.559402 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.559462 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.559479 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.559505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.559521 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.592107 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:03 crc kubenswrapper[4946]: E1203 06:51:03.592308 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.662789 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.662826 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.662835 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.662851 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.662859 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.765792 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.765858 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.765878 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.765901 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.765916 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.869477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.869512 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.869521 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.869537 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.869548 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.971599 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.971633 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.971641 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.971655 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:03 crc kubenswrapper[4946]: I1203 06:51:03.971665 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:03Z","lastTransitionTime":"2025-12-03T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.075391 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.075439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.075455 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.075490 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.075527 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.178941 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.178986 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.179000 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.179017 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.179029 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.282027 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.282057 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.282066 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.282079 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.282088 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.384907 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.384989 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.385013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.385045 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.385070 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.487928 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.487966 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.487974 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.487987 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.487996 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.591724 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.591810 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.591724 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.591980 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: E1203 06:51:04.592019 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.592048 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.592080 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.592116 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: E1203 06:51:04.592042 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:04 crc kubenswrapper[4946]: E1203 06:51:04.592159 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.592151 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.738480 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.738526 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.738535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.738552 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.738561 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.841152 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.841227 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.841252 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.841279 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.841297 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.944219 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.944272 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.944289 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.944312 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:04 crc kubenswrapper[4946]: I1203 06:51:04.944330 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:04Z","lastTransitionTime":"2025-12-03T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.046835 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.046865 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.046874 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.046888 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.046897 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.149728 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.149818 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.149834 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.149857 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.149874 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.253032 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.253106 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.253125 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.253154 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.253171 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.356655 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.356780 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.356806 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.356841 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.356864 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.459211 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.459259 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.459303 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.459320 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.459332 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.562461 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.562512 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.562527 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.562551 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.562565 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.592038 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:05 crc kubenswrapper[4946]: E1203 06:51:05.592164 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.665246 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.665301 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.665317 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.665338 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.665353 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.767596 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.768213 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.768343 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.768459 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.768556 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.871191 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.871262 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.871280 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.871303 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.871320 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.973865 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.973927 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.973944 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.973968 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:05 crc kubenswrapper[4946]: I1203 06:51:05.973985 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:05Z","lastTransitionTime":"2025-12-03T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.077074 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.077164 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.077183 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.077238 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.077255 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.180065 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.180139 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.180161 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.180210 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.180233 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.282848 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.282895 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.282906 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.282923 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.282935 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.385553 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.385625 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.385647 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.385677 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.385702 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.488726 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.488829 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.488848 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.488875 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.488893 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.591528 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.591578 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.591598 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.591623 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.591641 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.591771 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:06 crc kubenswrapper[4946]: E1203 06:51:06.591943 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.591997 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.592057 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:06 crc kubenswrapper[4946]: E1203 06:51:06.592164 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:06 crc kubenswrapper[4946]: E1203 06:51:06.592286 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.694258 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.694302 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.694311 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.694327 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.694353 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.796933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.797001 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.797018 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.797042 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.797062 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.900151 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.900234 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.900254 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.900278 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:06 crc kubenswrapper[4946]: I1203 06:51:06.900296 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:06Z","lastTransitionTime":"2025-12-03T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.003708 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.003806 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.003825 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.003852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.003872 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.107380 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.107462 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.107481 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.107507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.107534 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.211028 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.211089 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.211110 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.211139 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.211159 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.314087 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.314163 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.314184 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.314212 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.314229 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.417162 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.417235 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.417259 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.417293 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.417315 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.519764 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.519823 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.519841 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.519867 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.519884 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.592444 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:07 crc kubenswrapper[4946]: E1203 06:51:07.593779 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.607810 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.623492 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.623529 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.623548 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.623566 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.623579 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.629117 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.641628 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.651091 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.667287 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.683599 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.699706 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.717396 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.725772 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.725801 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.725812 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.725829 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.725876 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.737563 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.750971 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.764303 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.776974 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.791719 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.809927 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.826504 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.828833 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.828894 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.828910 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.828929 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.828941 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.844541 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.867733 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.881184 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:07Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.930559 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.930628 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.930649 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.930675 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:07 crc kubenswrapper[4946]: I1203 06:51:07.930696 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:07Z","lastTransitionTime":"2025-12-03T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.034335 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.034391 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.034408 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.034432 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.034448 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.137974 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.138086 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.138112 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.138142 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.138165 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.244584 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.244642 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.244660 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.244685 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.244702 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.348112 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.348380 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.348439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.348507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.348571 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.450942 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.450983 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.450995 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.451012 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.451023 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.553242 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.553288 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.553297 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.553315 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.553324 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.592383 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.592433 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.592546 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:08 crc kubenswrapper[4946]: E1203 06:51:08.592671 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:08 crc kubenswrapper[4946]: E1203 06:51:08.592898 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:08 crc kubenswrapper[4946]: E1203 06:51:08.592958 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.656802 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.656879 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.656902 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.656978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.657000 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.760358 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.760416 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.760432 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.760462 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.760479 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.862539 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.862574 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.862583 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.862597 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.862606 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.965834 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.965900 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.965912 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.965925 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:08 crc kubenswrapper[4946]: I1203 06:51:08.965943 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:08Z","lastTransitionTime":"2025-12-03T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.068812 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.068877 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.068897 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.068925 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.068945 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.171781 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.171822 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.171832 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.171848 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.171859 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.275636 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.275687 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.275731 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.275795 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.275812 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.379161 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.379196 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.379205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.379217 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.379225 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.482581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.482640 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.482658 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.482682 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.482700 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.585698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.585868 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.585901 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.585931 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.585952 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.592003 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:09 crc kubenswrapper[4946]: E1203 06:51:09.592198 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.689188 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.689253 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.689277 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.689302 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.689322 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.791376 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.791419 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.791430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.791447 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.791458 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.894512 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.894564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.894581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.894604 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:09 crc kubenswrapper[4946]: I1203 06:51:09.894621 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:09Z","lastTransitionTime":"2025-12-03T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.000406 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.000482 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.000507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.000539 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.000564 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.032978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.033038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.033056 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.033081 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.033099 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.055931 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:10Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.062420 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.062492 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.062517 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.062552 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.062576 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.082187 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:10Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.087526 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.087595 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.087674 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.087699 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.087832 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.103810 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:10Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.109969 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.110015 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.110035 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.110060 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.110082 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.128911 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:10Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.134875 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.134919 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.134936 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.134958 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.134976 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.156550 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:10Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.156711 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.158544 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.158569 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.158601 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.158614 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.158623 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.266725 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.267391 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.267403 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.267419 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.267429 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.369804 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.369844 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.369856 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.369874 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.369886 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.472703 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.472778 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.472796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.472818 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.472832 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.575855 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.575901 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.575909 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.575926 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.575937 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.592274 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.592492 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.592653 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.592705 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.592939 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:10 crc kubenswrapper[4946]: E1203 06:51:10.593018 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.679603 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.679642 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.679657 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.679679 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.679697 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.782458 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.782519 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.782535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.782561 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.782578 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.884896 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.884983 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.884998 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.885023 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.885040 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.987962 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.988008 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.988021 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.988044 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:10 crc kubenswrapper[4946]: I1203 06:51:10.988067 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:10Z","lastTransitionTime":"2025-12-03T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.090838 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.090915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.090933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.090960 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.090979 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.193277 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.193320 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.193337 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.193354 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.193367 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.296215 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.296261 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.296270 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.296285 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.296296 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.399235 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.399286 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.399298 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.399320 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.399334 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.502375 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.502477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.502497 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.502520 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.502539 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.591915 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:11 crc kubenswrapper[4946]: E1203 06:51:11.592116 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.605382 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.605428 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.605459 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.605481 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.605501 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.707405 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.707449 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.707460 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.707480 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.707492 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.810019 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.810099 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.810122 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.810155 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.810177 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.912276 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.912355 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.912389 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.912417 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:11 crc kubenswrapper[4946]: I1203 06:51:11.912437 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:11Z","lastTransitionTime":"2025-12-03T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.015005 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.015060 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.015073 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.015095 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.015108 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.118791 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.118834 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.118843 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.118858 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.118867 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.221994 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.222044 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.222061 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.222084 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.222099 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.324599 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.324628 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.324636 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.324651 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.324661 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.426479 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.426573 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.426599 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.426636 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.426660 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.529835 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.529909 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.529927 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.529953 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.529968 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.592448 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:12 crc kubenswrapper[4946]: E1203 06:51:12.592852 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.592671 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:12 crc kubenswrapper[4946]: E1203 06:51:12.593066 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.592448 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:12 crc kubenswrapper[4946]: E1203 06:51:12.593252 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.633049 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.633320 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.633411 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.633495 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.633582 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.736665 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.736974 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.737081 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.737170 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.737271 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.844146 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.844192 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.844203 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.844221 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.844235 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.947021 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.947384 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.947578 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.947820 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:12 crc kubenswrapper[4946]: I1203 06:51:12.948056 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:12Z","lastTransitionTime":"2025-12-03T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.051394 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.052271 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.052388 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.052505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.052624 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.155624 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.155668 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.155681 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.155699 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.155711 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.258410 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.258461 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.258474 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.258495 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.258511 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.360877 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.360943 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.360966 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.360993 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.361012 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.464477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.464522 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.464532 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.464548 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.464559 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.567668 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.567712 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.567723 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.567763 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.567775 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.592064 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:13 crc kubenswrapper[4946]: E1203 06:51:13.592196 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.670042 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.670330 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.670430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.670520 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.670595 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.773558 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.773635 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.773659 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.773691 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.773713 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.876544 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.876585 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.876598 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.876616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.876627 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.979521 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.979557 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.979565 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.979580 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:13 crc kubenswrapper[4946]: I1203 06:51:13.979589 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:13Z","lastTransitionTime":"2025-12-03T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.082143 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.082194 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.082204 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.082221 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.082232 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.184807 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.184879 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.184904 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.184929 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.184944 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.287614 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.287679 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.287690 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.287708 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.287718 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.390092 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.390128 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.390139 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.390155 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.390165 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.493676 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.493773 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.493793 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.493830 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.493864 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.591881 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.591923 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:14 crc kubenswrapper[4946]: E1203 06:51:14.592019 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.591900 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:14 crc kubenswrapper[4946]: E1203 06:51:14.592338 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:14 crc kubenswrapper[4946]: E1203 06:51:14.592505 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.592698 4946 scope.go:117] "RemoveContainer" containerID="ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e" Dec 03 06:51:14 crc kubenswrapper[4946]: E1203 06:51:14.592990 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.597438 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.597501 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.597521 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.597547 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.597564 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.699916 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.699982 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.700001 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.700027 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.700045 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.802973 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.803024 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.803033 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.803051 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.803061 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.905681 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.905726 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.905755 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.905773 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:14 crc kubenswrapper[4946]: I1203 06:51:14.905784 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:14Z","lastTransitionTime":"2025-12-03T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.009364 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.009416 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.009428 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.009446 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.009458 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.113429 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.113496 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.113513 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.113539 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.113557 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.217279 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.217382 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.217395 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.217414 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.217426 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.320534 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.320582 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.320594 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.320616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.320628 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.423219 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.423288 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.423305 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.423332 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.423351 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.525978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.526040 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.526053 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.526072 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.526084 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.592246 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:15 crc kubenswrapper[4946]: E1203 06:51:15.592410 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.606314 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.628343 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.628477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.628544 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.628607 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.628675 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.731578 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.731633 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.731650 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.731666 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.731678 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.834825 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.834875 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.834886 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.834904 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.834914 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.937958 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.938020 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.938039 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.938064 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:15 crc kubenswrapper[4946]: I1203 06:51:15.938082 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:15Z","lastTransitionTime":"2025-12-03T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.040205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.040261 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.040278 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.040299 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.040316 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.142247 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.142281 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.142290 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.142305 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.142383 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.244622 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.244662 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.244673 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.244688 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.244700 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.347621 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.347668 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.347678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.347698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.347714 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.450501 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.450568 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.450593 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.450626 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.450649 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.553204 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.553278 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.553289 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.553311 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.553325 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.592023 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.592128 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.592048 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:16 crc kubenswrapper[4946]: E1203 06:51:16.592239 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:16 crc kubenswrapper[4946]: E1203 06:51:16.592353 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:16 crc kubenswrapper[4946]: E1203 06:51:16.592570 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.655863 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.655907 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.655919 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.655938 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.655950 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.758318 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.758396 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.758420 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.758449 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.758470 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.861543 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.861622 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.861640 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.861666 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.861799 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.964573 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.964632 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.964651 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.964678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:16 crc kubenswrapper[4946]: I1203 06:51:16.964695 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:16Z","lastTransitionTime":"2025-12-03T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.072421 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.072470 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.072494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.072518 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.072534 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.175663 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.175709 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.175721 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.175774 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.175790 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.278944 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.279015 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.279040 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.279074 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.279095 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.382510 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.382566 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.382579 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.382899 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.382952 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.486045 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.486113 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.486136 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.486165 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.486186 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.588545 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.588595 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.588610 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.588631 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.588648 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.592184 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:17 crc kubenswrapper[4946]: E1203 06:51:17.592339 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.608343 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.625555 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.639527 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.656335 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.675334 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.690542 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.690612 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.690636 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.690662 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.690683 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.692453 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.712531 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.723993 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.732101 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.742823 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.758574 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.768051 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.776765 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.826861 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.826912 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.826925 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.826944 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.826959 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.827475 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.836817 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.849817 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.861143 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.878775 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.890307 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:17Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.930148 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.930196 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.930205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.930220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:17 crc kubenswrapper[4946]: I1203 06:51:17.930231 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:17Z","lastTransitionTime":"2025-12-03T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.033105 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.033197 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.033222 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.033257 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.033281 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.135643 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.135814 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.135825 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.135848 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.135863 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.238858 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.238902 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.238913 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.238930 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.238942 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.341084 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.341122 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.341131 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.341146 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.341155 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.444449 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.444502 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.444515 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.444535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.444548 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.547450 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.547486 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.547494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.547528 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.547539 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.592521 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.592573 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:18 crc kubenswrapper[4946]: E1203 06:51:18.592637 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.592573 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:18 crc kubenswrapper[4946]: E1203 06:51:18.592810 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:18 crc kubenswrapper[4946]: E1203 06:51:18.592802 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.650529 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.650591 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.650608 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.650633 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.650650 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.753451 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.753502 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.753513 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.753533 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.753547 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.856489 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.856540 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.856551 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.856568 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.856580 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.961347 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.961384 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.961392 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.961409 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:18 crc kubenswrapper[4946]: I1203 06:51:18.961419 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:18Z","lastTransitionTime":"2025-12-03T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.063915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.063949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.063962 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.063978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.063990 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.166680 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.166715 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.166726 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.166766 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.166778 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.269771 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.269807 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.269816 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.269833 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.269843 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.372319 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.372355 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.372365 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.372384 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.372395 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.475769 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.475804 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.475815 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.475834 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.475845 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.578885 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.578933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.578949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.578973 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.578990 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.596050 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:19 crc kubenswrapper[4946]: E1203 06:51:19.596197 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.681509 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.681543 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.681553 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.681569 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.681580 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.771214 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:19 crc kubenswrapper[4946]: E1203 06:51:19.771475 4946 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:51:19 crc kubenswrapper[4946]: E1203 06:51:19.779055 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs podName:2e5abff8-4252-4371-803d-d241c81c5910 nodeName:}" failed. No retries permitted until 2025-12-03 06:51:51.778987658 +0000 UTC m=+104.575677837 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs") pod "network-metrics-daemon-hkpdf" (UID: "2e5abff8-4252-4371-803d-d241c81c5910") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.783680 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.783930 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.783952 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.783981 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.783993 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.887677 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.887722 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.887732 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.887767 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.887779 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.991411 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.991477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.991494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.991523 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:19 crc kubenswrapper[4946]: I1203 06:51:19.991542 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:19Z","lastTransitionTime":"2025-12-03T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.094824 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.094906 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.094927 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.094957 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.094983 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.197855 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.197902 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.197913 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.197930 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.197942 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.223579 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.223665 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.223678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.223700 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.223715 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.239121 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.243891 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.243955 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.243967 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.243987 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.244003 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.261798 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.267052 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.267099 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.267109 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.267128 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.267143 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.279207 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.283999 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.284212 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.284364 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.284507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.284657 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.298565 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.302985 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.303038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.303048 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.303064 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.303075 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.316179 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.316299 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.318570 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.318602 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.318615 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.318647 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.318663 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.421517 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.421556 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.421571 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.421603 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.421615 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.524377 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.524419 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.524432 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.524451 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.524463 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.592049 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.592118 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.592132 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.592193 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.592325 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:20 crc kubenswrapper[4946]: E1203 06:51:20.592401 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.626889 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.626995 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.627013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.627038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.627057 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.730155 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.730208 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.730220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.730239 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.730255 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.788788 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/0.log" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.788861 4946 generic.go:334] "Generic (PLEG): container finished" podID="08074f18-fe84-4d7b-8327-9696cbe78f38" containerID="79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7" exitCode=1 Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.788903 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t9hvz" event={"ID":"08074f18-fe84-4d7b-8327-9696cbe78f38","Type":"ContainerDied","Data":"79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.789343 4946 scope.go:117] "RemoveContainer" containerID="79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.810144 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.824379 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.834810 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.834857 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.834866 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.834881 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.834891 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.841187 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.870156 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.885797 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.897963 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.911171 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.931289 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.938817 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.938898 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.938916 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.938940 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.938981 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:20Z","lastTransitionTime":"2025-12-03T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.943853 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.955554 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.974936 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:20 crc kubenswrapper[4946]: I1203 06:51:20.993956 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:20Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.007115 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.020292 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.040201 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.041623 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.041708 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.041731 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.041880 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.041899 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.058091 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.074308 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.091572 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:19Z\\\",\\\"message\\\":\\\"2025-12-03T06:50:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e\\\\n2025-12-03T06:50:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e to /host/opt/cni/bin/\\\\n2025-12-03T06:50:34Z [verbose] multus-daemon started\\\\n2025-12-03T06:50:34Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:51:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.105628 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.148405 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.148466 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.148483 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.148510 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.148527 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.256099 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.256146 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.256159 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.256180 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.256194 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.359813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.359878 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.359891 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.359908 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.359920 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.464248 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.464395 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.464414 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.464491 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.464511 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.567357 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.567388 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.567396 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.567409 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.567417 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.591843 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:21 crc kubenswrapper[4946]: E1203 06:51:21.591987 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.670473 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.670516 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.670527 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.670542 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.670552 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.772986 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.773261 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.773329 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.773397 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.773460 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.793051 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/0.log" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.793108 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t9hvz" event={"ID":"08074f18-fe84-4d7b-8327-9696cbe78f38","Type":"ContainerStarted","Data":"e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.809850 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.830272 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.848903 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.868623 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.877889 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.878212 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.878388 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.878589 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.878721 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.889363 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:19Z\\\",\\\"message\\\":\\\"2025-12-03T06:50:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e\\\\n2025-12-03T06:50:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e to /host/opt/cni/bin/\\\\n2025-12-03T06:50:34Z [verbose] multus-daemon started\\\\n2025-12-03T06:50:34Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:51:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:51:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.904070 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.928905 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.946796 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.961244 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.982137 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.982203 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.982221 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.982246 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.982264 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:21Z","lastTransitionTime":"2025-12-03T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:21 crc kubenswrapper[4946]: I1203 06:51:21.987227 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.000097 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:21Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.011872 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.025776 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.043692 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.060994 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.073214 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.084838 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.084880 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.084890 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.084908 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.084919 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.088792 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.100528 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.121461 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.187103 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.187140 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.187150 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.187166 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.187177 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.289410 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.289471 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.289490 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.289514 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.289532 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.392439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.392505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.392527 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.392554 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.392571 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.496146 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.496214 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.496237 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.496265 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.496288 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.592215 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.592254 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.592287 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:22 crc kubenswrapper[4946]: E1203 06:51:22.592375 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:22 crc kubenswrapper[4946]: E1203 06:51:22.592528 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:22 crc kubenswrapper[4946]: E1203 06:51:22.592587 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.598252 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.598284 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.598293 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.598307 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.598317 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.701034 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.701086 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.701099 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.701118 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.701130 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.804473 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.804547 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.804573 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.804601 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.804622 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.906847 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.906905 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.906947 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.906970 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:22 crc kubenswrapper[4946]: I1203 06:51:22.906988 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:22Z","lastTransitionTime":"2025-12-03T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.009325 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.009378 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.009396 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.009418 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.009435 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.112199 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.112272 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.112289 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.112313 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.112334 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.215020 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.215072 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.215089 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.215117 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.215135 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.317574 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.317621 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.317631 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.317652 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.317663 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.420454 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.420526 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.420551 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.420581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.420604 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.523429 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.523475 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.523493 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.523514 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.523530 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.592161 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:23 crc kubenswrapper[4946]: E1203 06:51:23.592377 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.626702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.626781 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.626800 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.626824 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.626842 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.729598 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.729652 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.729678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.729706 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.729727 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.832751 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.832796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.832806 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.832824 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.832837 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.935499 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.935564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.935590 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.935619 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:23 crc kubenswrapper[4946]: I1203 06:51:23.935638 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:23Z","lastTransitionTime":"2025-12-03T06:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.039395 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.039459 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.039488 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.039520 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.039543 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.143198 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.143260 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.143276 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.143299 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.143315 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.247030 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.247115 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.247140 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.247168 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.247184 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.350023 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.350052 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.350060 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.350074 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.350082 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.453376 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.453417 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.453426 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.453469 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.453479 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.555711 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.555767 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.555801 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.555829 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.555842 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.592146 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.592192 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.592383 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:24 crc kubenswrapper[4946]: E1203 06:51:24.592676 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:24 crc kubenswrapper[4946]: E1203 06:51:24.593270 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:24 crc kubenswrapper[4946]: E1203 06:51:24.593170 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.659683 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.659771 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.659790 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.659814 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.659833 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.762555 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.762591 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.762599 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.762616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.762625 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.865772 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.865810 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.865822 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.865842 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.865857 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.969271 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.969315 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.969327 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.969347 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:24 crc kubenswrapper[4946]: I1203 06:51:24.969361 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:24Z","lastTransitionTime":"2025-12-03T06:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.072675 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.072733 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.072779 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.072804 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.072821 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.176430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.176714 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.176817 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.176910 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.176982 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.281027 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.281093 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.281111 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.281137 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.281155 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.383978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.384057 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.384081 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.384111 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.384132 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.487404 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.487461 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.487483 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.487507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.487523 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.591264 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.591808 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.591979 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.592191 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.592346 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.592042 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:25 crc kubenswrapper[4946]: E1203 06:51:25.592637 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.695209 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.695292 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.695324 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.695364 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.695390 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.798157 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.798218 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.798235 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.798256 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.798270 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.900929 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.901285 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.901499 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.901653 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:25 crc kubenswrapper[4946]: I1203 06:51:25.901837 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:25Z","lastTransitionTime":"2025-12-03T06:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.004701 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.005170 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.005361 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.005578 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.005776 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.109433 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.109490 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.109510 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.109535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.109554 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.212772 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.212823 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.212836 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.212857 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.212874 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.316484 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.316563 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.316613 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.316682 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.316709 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.420183 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.420247 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.420261 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.420282 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.420297 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.522812 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.522849 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.522859 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.522875 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.522884 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.592814 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.592926 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.592841 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:26 crc kubenswrapper[4946]: E1203 06:51:26.593071 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:26 crc kubenswrapper[4946]: E1203 06:51:26.593176 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:26 crc kubenswrapper[4946]: E1203 06:51:26.593354 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.626333 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.626385 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.626396 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.626414 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.626426 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.730475 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.730525 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.730536 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.730554 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.730568 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.833482 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.833536 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.833551 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.833574 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.833587 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.936624 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.936688 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.936706 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.936774 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:26 crc kubenswrapper[4946]: I1203 06:51:26.936797 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:26Z","lastTransitionTime":"2025-12-03T06:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.039969 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.040022 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.040043 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.040066 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.040174 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.143345 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.143413 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.143430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.143459 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.143480 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.246720 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.246830 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.246868 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.246903 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.246929 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.351223 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.351314 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.351338 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.351400 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.351422 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.454867 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.454928 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.454945 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.454970 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.454988 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.558061 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.558128 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.558145 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.558173 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.558196 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.591971 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:27 crc kubenswrapper[4946]: E1203 06:51:27.592375 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.593854 4946 scope.go:117] "RemoveContainer" containerID="ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.613652 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.633262 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.650357 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.660616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.660660 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.660674 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.660692 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.660705 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.660774 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.690270 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.707763 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.737658 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.751582 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.763792 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.763857 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.763876 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.763902 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.763919 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.764695 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.778453 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:19Z\\\",\\\"message\\\":\\\"2025-12-03T06:50:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e\\\\n2025-12-03T06:50:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e to /host/opt/cni/bin/\\\\n2025-12-03T06:50:34Z [verbose] multus-daemon started\\\\n2025-12-03T06:50:34Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:51:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:51:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.794190 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.813589 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.819943 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/2.log" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.839110 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.842521 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.843208 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.854983 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.874730 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.874791 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.874806 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.874845 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.874860 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.877556 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.892485 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.907185 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.922131 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.937416 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.958054 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.977530 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.977949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.977995 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.978011 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.978030 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.978042 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:27Z","lastTransitionTime":"2025-12-03T06:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:27 crc kubenswrapper[4946]: I1203 06:51:27.994829 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:27Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.013291 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.028090 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.045949 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.065348 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.080451 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.080500 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.080516 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.080542 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.080560 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.083907 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:19Z\\\",\\\"message\\\":\\\"2025-12-03T06:50:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e\\\\n2025-12-03T06:50:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e to /host/opt/cni/bin/\\\\n2025-12-03T06:50:34Z [verbose] multus-daemon started\\\\n2025-12-03T06:50:34Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:51:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:51:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.102972 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.117898 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.136421 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.149141 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.162793 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.172145 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.182344 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.183399 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.183433 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.183445 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.183464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.183478 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.193398 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.210244 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.224805 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.236255 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.285272 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.285315 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.285323 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.285337 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.285350 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.388257 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.388311 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.388325 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.388347 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.388356 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.492014 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.492072 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.492091 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.492114 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.492131 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.592489 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.592646 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:28 crc kubenswrapper[4946]: E1203 06:51:28.592891 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.592941 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:28 crc kubenswrapper[4946]: E1203 06:51:28.593117 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:28 crc kubenswrapper[4946]: E1203 06:51:28.593338 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.600128 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.600205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.600231 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.600264 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.600298 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.703293 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.703341 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.703352 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.703366 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.703375 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.806347 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.806394 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.806406 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.806422 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.806432 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.908685 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.908821 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.908844 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.908871 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:28 crc kubenswrapper[4946]: I1203 06:51:28.908890 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:28Z","lastTransitionTime":"2025-12-03T06:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.012465 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.012548 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.012571 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.012603 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.012625 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.115359 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.115409 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.115425 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.115448 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.115465 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.219107 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.219157 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.219173 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.219193 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.219207 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.324065 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.324344 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.324404 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.324487 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.324556 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.427851 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.427923 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.427940 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.427969 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.427988 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.531405 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.531801 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.531942 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.532076 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.532217 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.592838 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:29 crc kubenswrapper[4946]: E1203 06:51:29.592975 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.635442 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.635494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.635505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.635529 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.635543 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.738852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.738950 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.738982 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.739020 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.739044 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.842540 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.842601 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.842619 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.842646 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.842665 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.853081 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/3.log" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.854109 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/2.log" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.857559 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" exitCode=1 Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.857608 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.857642 4946 scope.go:117] "RemoveContainer" containerID="ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.858851 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 06:51:29 crc kubenswrapper[4946]: E1203 06:51:29.859108 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.896110 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.917378 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.941777 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.946246 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.946305 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.946329 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.946358 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.946380 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:29Z","lastTransitionTime":"2025-12-03T06:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.966004 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:29 crc kubenswrapper[4946]: I1203 06:51:29.985411 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:29Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.004909 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.022210 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.036409 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.049205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.049266 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.049283 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.049310 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.049331 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.058578 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:19Z\\\",\\\"message\\\":\\\"2025-12-03T06:50:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e\\\\n2025-12-03T06:50:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e to /host/opt/cni/bin/\\\\n2025-12-03T06:50:34Z [verbose] multus-daemon started\\\\n2025-12-03T06:50:34Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:51:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:51:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.075012 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.092923 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.111592 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.133472 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.153077 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.153137 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.153153 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.153176 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.153195 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.159828 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.172987 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.188036 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.200767 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.224223 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:28Z\\\",\\\"message\\\":\\\"ndler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 06:51:28.537265 6953 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.235873 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.256007 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.256051 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.256069 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.256092 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.256111 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.358941 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.359020 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.359042 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.359075 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.359100 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.400869 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.401115 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:34.401067482 +0000 UTC m=+147.197757631 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.462480 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.462522 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.462533 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.462550 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.462561 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.502603 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.502660 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.502681 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.502709 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.502798 4946 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.502843 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:52:34.502831142 +0000 UTC m=+147.299521251 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.502873 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.502912 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.502926 4946 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.502986 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:52:34.502966916 +0000 UTC m=+147.299657075 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.503017 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.503066 4946 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.503087 4946 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.503181 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:52:34.503156661 +0000 UTC m=+147.299846810 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.503258 4946 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.503305 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:52:34.503294455 +0000 UTC m=+147.299984664 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.565809 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.565858 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.565872 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.565890 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.565902 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.592438 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.592522 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.592595 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.592677 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.592460 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.592877 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.668399 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.668451 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.668474 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.668505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.668527 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.687397 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.687463 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.687485 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.687515 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.687538 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.706626 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.711633 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.711696 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.711717 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.711802 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.711830 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.731591 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.738981 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.739214 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.739238 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.739262 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.739280 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.760604 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.765906 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.765953 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.765963 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.765982 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.765998 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.783532 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.788392 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.788448 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.788464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.788496 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.788513 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.807528 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:30Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:30 crc kubenswrapper[4946]: E1203 06:51:30.807808 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.809720 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.809781 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.809793 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.809811 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.809823 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.864269 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/3.log" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.912933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.912997 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.913019 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.913050 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:30 crc kubenswrapper[4946]: I1203 06:51:30.913074 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:30Z","lastTransitionTime":"2025-12-03T06:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.015489 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.015554 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.015570 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.015594 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.015614 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.118672 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.118734 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.118779 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.118805 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.118822 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.222156 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.222224 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.222236 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.222255 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.222267 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.332384 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.332441 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.332454 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.332472 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.332484 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.434684 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.434726 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.434762 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.434783 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.434797 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.537352 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.537407 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.537424 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.537446 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.537464 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.592403 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:31 crc kubenswrapper[4946]: E1203 06:51:31.592596 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.639934 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.639982 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.639995 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.640013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.640025 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.743685 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.743726 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.743751 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.743768 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.743780 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.846409 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.846439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.846448 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.846461 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.846469 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.950191 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.950235 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.950244 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.950261 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:31 crc kubenswrapper[4946]: I1203 06:51:31.950270 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:31Z","lastTransitionTime":"2025-12-03T06:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.053693 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.053791 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.053813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.053841 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.053855 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.156793 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.156840 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.156849 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.156865 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.156876 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.260168 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.260204 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.260214 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.260228 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.260236 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.363690 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.363807 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.363828 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.363854 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.363872 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.466618 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.466667 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.466681 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.466701 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.466714 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.570577 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.570646 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.570665 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.570691 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.570710 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.592381 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.592478 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.592577 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:32 crc kubenswrapper[4946]: E1203 06:51:32.592867 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:32 crc kubenswrapper[4946]: E1203 06:51:32.592990 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:32 crc kubenswrapper[4946]: E1203 06:51:32.593097 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.673617 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.673674 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.673698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.673721 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.673755 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.777676 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.777757 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.777771 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.777791 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.777802 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.880872 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.880907 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.880918 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.880934 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.880947 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.983859 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.983926 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.983944 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.983969 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:32 crc kubenswrapper[4946]: I1203 06:51:32.983986 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:32Z","lastTransitionTime":"2025-12-03T06:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.086707 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.086766 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.086778 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.086797 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.086810 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.190176 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.190265 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.190297 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.190327 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.190348 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.292933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.292999 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.293019 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.293049 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.293067 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.395704 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.395807 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.395825 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.395852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.395869 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.499285 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.499334 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.499345 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.499362 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.499381 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.592617 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:33 crc kubenswrapper[4946]: E1203 06:51:33.592795 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.602180 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.602215 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.602225 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.602236 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.602245 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.712335 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.712393 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.712411 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.712435 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.712452 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.817034 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.817070 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.817079 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.817097 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.817107 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.919788 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.919830 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.919843 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.919864 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:33 crc kubenswrapper[4946]: I1203 06:51:33.919874 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:33Z","lastTransitionTime":"2025-12-03T06:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.022916 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.023001 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.023013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.023034 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.023069 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.126521 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.126589 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.126607 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.126635 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.126651 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.229441 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.229493 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.229503 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.229522 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.229535 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.332204 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.332252 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.332262 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.332283 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.332294 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.435956 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.436014 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.436026 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.436049 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.436064 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.538603 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.538644 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.538657 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.538677 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.538690 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.592628 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.592628 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:34 crc kubenswrapper[4946]: E1203 06:51:34.592833 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.592659 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:34 crc kubenswrapper[4946]: E1203 06:51:34.592942 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:34 crc kubenswrapper[4946]: E1203 06:51:34.593007 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.641650 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.641716 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.641740 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.641807 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.641823 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.744588 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.744676 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.744698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.744723 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.744789 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.848469 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.848535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.848553 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.848924 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.848964 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.952128 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.952546 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.952646 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.952771 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:34 crc kubenswrapper[4946]: I1203 06:51:34.952867 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:34Z","lastTransitionTime":"2025-12-03T06:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.056452 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.056986 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.057071 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.057145 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.057213 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.160942 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.161011 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.161030 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.161058 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.161082 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.264575 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.264621 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.264632 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.264649 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.264660 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.368597 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.368696 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.368717 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.368806 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.368828 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.471385 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.471430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.471439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.471454 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.471470 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.574576 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.574702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.574816 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.574915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.574946 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.591890 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:35 crc kubenswrapper[4946]: E1203 06:51:35.592102 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.678052 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.678084 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.678092 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.678106 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.678115 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.781151 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.781198 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.781212 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.781232 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.781330 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.884335 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.884422 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.884448 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.884483 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.884505 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.986867 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.986938 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.986953 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.986978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:35 crc kubenswrapper[4946]: I1203 06:51:35.986995 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:35Z","lastTransitionTime":"2025-12-03T06:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.090272 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.090340 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.090351 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.090374 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.090386 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.194626 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.194686 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.194698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.194718 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.194732 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.297612 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.297651 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.297661 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.297675 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.297686 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.399684 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.399727 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.399754 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.399771 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.399784 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.503078 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.503142 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.503159 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.503184 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.503202 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.592834 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.592919 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.592923 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:36 crc kubenswrapper[4946]: E1203 06:51:36.593406 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:36 crc kubenswrapper[4946]: E1203 06:51:36.593802 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:36 crc kubenswrapper[4946]: E1203 06:51:36.593919 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.605971 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.606043 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.606137 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.606169 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.606192 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.709702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.709807 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.709828 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.709854 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.709872 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.813395 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.813448 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.813464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.813489 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.813506 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.916654 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.916719 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.916775 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.916807 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:36 crc kubenswrapper[4946]: I1203 06:51:36.916831 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:36Z","lastTransitionTime":"2025-12-03T06:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.020820 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.020894 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.020915 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.020945 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.020968 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.123551 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.123581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.123592 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.123608 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.123620 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.226545 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.226616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.226635 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.226659 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.226675 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.329702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.329796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.329816 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.329833 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.329844 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.432769 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.433180 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.433198 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.433223 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.433243 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.536681 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.536864 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.536883 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.536907 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.536926 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.592731 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:37 crc kubenswrapper[4946]: E1203 06:51:37.593055 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.630706 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.640848 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.640896 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.640911 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.640930 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.640945 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.655204 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.671314 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.692248 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.713654 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.726499 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.740188 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.743507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.743550 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.743564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.743587 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.743603 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.755616 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:19Z\\\",\\\"message\\\":\\\"2025-12-03T06:50:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e\\\\n2025-12-03T06:50:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e to /host/opt/cni/bin/\\\\n2025-12-03T06:50:34Z [verbose] multus-daemon started\\\\n2025-12-03T06:50:34Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:51:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:51:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.767609 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.780069 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.796450 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.809795 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.825950 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.845576 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.847392 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.847434 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.847443 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.847459 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.847469 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.861661 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.874700 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.896384 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea81657b6337ddbfafb897a92c721b399cfd2497f50a1d5d693fd8678e0f628e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:50:59Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI1203 06:50:59.488962 6572 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:50:59.489009 6572 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 06:50:59.489018 6572 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 06:50:59.489041 6572 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:50:59.489082 6572 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:50:59.489109 6572 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:50:59.489136 6572 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:50:59.489167 6572 factory.go:656] Stopping watch factory\\\\nI1203 06:50:59.489193 6572 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:50:59.489201 6572 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:50:59.489254 6572 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:50:59.489268 6572 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 06:50:59.489531 6572 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 06:50:59.489619 6572 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:50:59.489695 6572 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:50:59.489911 6572 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:28Z\\\",\\\"message\\\":\\\"ndler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 06:51:28.537265 6953 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.913192 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.923461 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:37Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.951003 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.951086 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.951113 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.951147 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:37 crc kubenswrapper[4946]: I1203 06:51:37.951170 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:37Z","lastTransitionTime":"2025-12-03T06:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.054540 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.054639 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.054656 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.054682 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.054700 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.157429 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.157482 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.157498 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.157524 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.157543 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.260409 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.260474 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.260497 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.260526 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.260549 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.364280 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.364347 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.364370 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.364398 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.364420 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.467734 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.467918 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.467944 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.468000 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.468019 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.570919 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.570989 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.571014 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.571043 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.571060 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.592720 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.592844 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.592721 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:38 crc kubenswrapper[4946]: E1203 06:51:38.592938 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:38 crc kubenswrapper[4946]: E1203 06:51:38.593084 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:38 crc kubenswrapper[4946]: E1203 06:51:38.593290 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.674853 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.674923 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.674944 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.674976 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.674998 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.778923 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.778986 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.779004 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.779031 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.779049 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.882081 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.882143 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.882162 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.882187 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.882217 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.985514 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.985574 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.985586 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.985607 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:38 crc kubenswrapper[4946]: I1203 06:51:38.985620 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:38Z","lastTransitionTime":"2025-12-03T06:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.089362 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.089416 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.089427 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.089448 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.089461 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.192622 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.192658 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.192666 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.192678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.192690 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.296136 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.296203 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.296220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.296248 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.296267 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.398932 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.399005 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.399028 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.399057 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.399080 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.501945 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.502017 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.502043 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.502076 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.502102 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.591889 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:39 crc kubenswrapper[4946]: E1203 06:51:39.592090 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.604402 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.604466 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.604483 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.604507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.604524 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.707436 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.707514 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.707537 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.707568 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.707590 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.811634 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.811707 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.811733 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.811805 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.811829 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.915103 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.915174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.915199 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.915230 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:39 crc kubenswrapper[4946]: I1203 06:51:39.915251 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:39Z","lastTransitionTime":"2025-12-03T06:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.018092 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.018145 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.018162 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.018185 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.018203 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.121499 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.121569 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.121591 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.121624 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.121645 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.225174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.225245 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.225267 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.225300 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.225324 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.328156 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.328207 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.328222 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.328246 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.328261 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.432511 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.432573 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.432590 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.432616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.432638 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.535683 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.535730 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.535775 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.535792 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.535801 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.591854 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.591903 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.591954 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:40 crc kubenswrapper[4946]: E1203 06:51:40.592129 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:40 crc kubenswrapper[4946]: E1203 06:51:40.592290 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:40 crc kubenswrapper[4946]: E1203 06:51:40.592514 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.638301 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.638368 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.638389 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.638414 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.638432 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.742305 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.742356 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.742372 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.742396 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.742424 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.846120 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.846270 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.846295 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.846326 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.846348 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.949280 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.949342 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.949359 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.949383 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.949403 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.966604 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.966659 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.966676 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.966700 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.966716 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:40 crc kubenswrapper[4946]: E1203 06:51:40.987410 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:40Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.992089 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.992134 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.992146 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.992163 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:40 crc kubenswrapper[4946]: I1203 06:51:40.992174 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:40Z","lastTransitionTime":"2025-12-03T06:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: E1203 06:51:41.008967 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.012632 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.012705 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.012729 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.012803 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.012826 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: E1203 06:51:41.031551 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.035085 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.035258 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.035347 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.035456 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.035549 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: E1203 06:51:41.053072 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.058028 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.058104 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.058129 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.058161 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.058186 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: E1203 06:51:41.078302 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:41 crc kubenswrapper[4946]: E1203 06:51:41.078487 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.080813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.080861 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.080880 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.080904 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.080919 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.184316 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.184371 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.184386 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.184408 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.184427 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.288072 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.288168 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.288190 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.288220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.288249 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.391507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.391582 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.391604 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.391630 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.391649 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.493629 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.493682 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.493694 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.493711 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.493723 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.592331 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:41 crc kubenswrapper[4946]: E1203 06:51:41.592489 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.596719 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.596796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.596811 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.596830 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.596842 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.700038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.700132 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.700150 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.700173 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.700189 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.803883 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.803937 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.803949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.803968 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.803980 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.907531 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.907630 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.907648 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.907703 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:41 crc kubenswrapper[4946]: I1203 06:51:41.907723 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:41Z","lastTransitionTime":"2025-12-03T06:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.010191 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.010282 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.010305 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.010326 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.010344 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.114117 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.114182 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.114199 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.114224 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.114241 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.216962 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.217007 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.217015 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.217033 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.217046 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.319845 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.319914 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.319932 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.319958 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.319977 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.423733 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.423821 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.423839 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.423869 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.423887 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.527816 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.527872 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.527888 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.527916 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.527933 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.591962 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.592035 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:42 crc kubenswrapper[4946]: E1203 06:51:42.592148 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:42 crc kubenswrapper[4946]: E1203 06:51:42.592283 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.592368 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:42 crc kubenswrapper[4946]: E1203 06:51:42.592537 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.631449 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.631506 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.631522 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.631538 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.631553 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.735109 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.735194 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.735219 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.735258 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.735282 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.838935 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.838992 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.839009 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.839033 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.839050 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.942763 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.942823 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.942837 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.942858 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:42 crc kubenswrapper[4946]: I1203 06:51:42.942871 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:42Z","lastTransitionTime":"2025-12-03T06:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.045883 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.045936 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.045952 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.045976 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.045996 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.150105 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.150172 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.150189 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.150580 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.150608 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.252436 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.252469 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.252476 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.252495 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.252506 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.355933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.356006 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.356030 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.356065 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.356087 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.459370 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.459480 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.459501 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.459528 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.459546 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.563382 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.563440 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.563453 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.563474 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.563488 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.592594 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:43 crc kubenswrapper[4946]: E1203 06:51:43.592794 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.666381 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.666443 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.666454 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.666473 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.666486 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.770218 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.770285 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.770296 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.770313 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.770326 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.873799 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.873842 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.873852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.873869 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.873880 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.977610 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.977715 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.977830 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.977926 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:43 crc kubenswrapper[4946]: I1203 06:51:43.977954 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:43Z","lastTransitionTime":"2025-12-03T06:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.080507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.080571 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.080588 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.080616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.080632 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.184055 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.184113 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.184130 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.184152 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.184169 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.286983 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.287077 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.287097 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.287119 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.287136 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.389993 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.390038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.390048 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.390114 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.390126 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.493351 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.493410 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.493427 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.493452 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.493471 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.592728 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.592856 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.592908 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:44 crc kubenswrapper[4946]: E1203 06:51:44.593084 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:44 crc kubenswrapper[4946]: E1203 06:51:44.593297 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:44 crc kubenswrapper[4946]: E1203 06:51:44.593867 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.594522 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 06:51:44 crc kubenswrapper[4946]: E1203 06:51:44.594884 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.597196 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.597310 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.597375 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.597400 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.597420 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.611046 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.625471 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.641395 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.664951 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.683462 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.699961 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.700036 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.700053 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.700080 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.700099 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.702238 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.717910 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.750798 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:28Z\\\",\\\"message\\\":\\\"ndler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 06:51:28.537265 6953 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:51:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.767614 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.783199 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.803459 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.803507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.803523 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.803545 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.803561 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.808180 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.822857 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.841482 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.861030 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.892467 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.905893 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.905940 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.905953 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.905973 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.905989 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:44Z","lastTransitionTime":"2025-12-03T06:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.911590 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.926761 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.941121 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:19Z\\\",\\\"message\\\":\\\"2025-12-03T06:50:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e\\\\n2025-12-03T06:50:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e to /host/opt/cni/bin/\\\\n2025-12-03T06:50:34Z [verbose] multus-daemon started\\\\n2025-12-03T06:50:34Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:51:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:51:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:44 crc kubenswrapper[4946]: I1203 06:51:44.953141 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:44Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.009300 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.009356 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.009371 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.009394 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.009411 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.112013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.112062 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.112078 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.112101 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.112121 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.214817 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.214865 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.214877 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.214894 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.214907 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.317727 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.317823 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.317840 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.317865 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.317881 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.421586 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.421639 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.421656 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.421679 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.421696 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.525055 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.525246 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.525269 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.525296 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.525317 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.592266 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:45 crc kubenswrapper[4946]: E1203 06:51:45.592508 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.627789 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.627852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.627868 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.627892 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.627910 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.730657 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.730733 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.730803 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.730835 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.730857 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.834107 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.834185 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.834205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.834232 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.834251 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.937052 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.937112 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.937131 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.937157 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:45 crc kubenswrapper[4946]: I1203 06:51:45.937174 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:45Z","lastTransitionTime":"2025-12-03T06:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.041213 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.041267 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.041283 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.041311 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.041328 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.144963 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.145026 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.145045 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.145070 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.145087 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.248372 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.248432 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.248450 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.248474 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.248492 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.351987 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.352087 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.352106 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.352150 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.352169 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.455429 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.455505 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.455514 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.455531 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.455542 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.558059 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.558136 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.558171 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.558203 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.558225 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.592258 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.592305 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.592258 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:46 crc kubenswrapper[4946]: E1203 06:51:46.592397 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:46 crc kubenswrapper[4946]: E1203 06:51:46.592663 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:46 crc kubenswrapper[4946]: E1203 06:51:46.592834 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.661814 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.661874 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.661893 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.661916 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.661934 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.769203 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.769252 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.769270 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.769288 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.769299 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.872170 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.872205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.872213 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.872226 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.872235 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.975957 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.976004 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.976014 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.976054 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:46 crc kubenswrapper[4946]: I1203 06:51:46.976065 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:46Z","lastTransitionTime":"2025-12-03T06:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.078764 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.078813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.078840 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.078859 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.078870 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.181968 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.182033 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.182047 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.182071 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.182088 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.286362 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.286424 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.286441 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.286464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.286481 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.388504 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.388559 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.388569 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.388587 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.388601 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.491013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.491082 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.491100 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.491127 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.491148 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.592217 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:47 crc kubenswrapper[4946]: E1203 06:51:47.592411 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.594712 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.594782 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.594799 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.594820 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.594840 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.651699 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.666892 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2b4cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"188b3f12-d66a-4447-979f-efea0e31abf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc5d1f3521b9fd492db2d0c39a4bd645e6da44c5e89f05635c7c131fa987d14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2sgrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2b4cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.693439 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58ed82b9-82e5-4d9a-b331-80c32397dc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:28Z\\\",\\\"message\\\":\\\"ndler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:28Z is after 2025-08-24T17:21:41Z]\\\\nI1203 06:51:28.537265 6953 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:51:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzcsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6fppr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.697068 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.697104 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.697115 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.697133 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.697146 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.709655 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4003d158-6bdd-45bd-a68c-ca52bd7264c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc100a4a612f971731b665793e93e310f2bd51b2d593818a0de9c86b792bef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pp9rm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6bt2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.724424 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8plsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84d71a3-4e42-48a6-802a-553364d32a9b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dedcfd9a57cea74c31ecddd75f0ed43808b758136a3c635df887b822c20f151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86mzf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8plsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.748976 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c97c8f0f-b4bc-44e9-aeae-cf5765f4fc78\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7aed287f526bfeff70ed6f2789a84dfdd98c4d6a2068e2c87ff7ab3618c89e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67baf0489e961b39fe28bf74644ecb902ef7723e4c22c5fe54e43657049629db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7779fd49f2231ac718db8a1ae938b64a623d5f57741d40cc592ce5bd91f2d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e778e533417a1d755402be6bcc2d9765e7ca34f91d5c835489aea209fd564969\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43a7fb5e8572fab3221000819693f9db8ac7b86d7a88f83b80a32d327ed1a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0accdc24752dd4f318ae929b53ab8abeb9741da1c18478dd23c702a7996bc3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffbf3e46c4de0fbfe3d5b1bb76c778a7b0c5079c738f4ab662183c0801a6c0e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d9a5f7e544934fb3b310974e6b8bd564846f4f9634b829a0f9586fc1bd29c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.765944 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.783988 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c68ed8199a58aa2444a33f76e52d96b193aeb9691e2fc9fe79442773228419f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.800949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.801021 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.801041 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.801067 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.801086 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.802000 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce18076cf29687997b94a6e8ccef0e5a3309f93566b554457dac1ef1bd418bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d921c89363e0ce7679935b67f13cc757c168ccd9356b7f2320f068dc75d425f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.817851 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68bb29de-3574-4bd9-aa64-1da58d3dd47e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:50:25Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:50:19.854278 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:50:19.855780 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3370659246/tls.crt::/tmp/serving-cert-3370659246/tls.key\\\\\\\"\\\\nI1203 06:50:25.557461 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:50:25.561639 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:50:25.561674 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:50:25.561784 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:50:25.561797 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:50:25.570991 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1203 06:50:25.571032 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:50:25.571044 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571059 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:50:25.571074 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:50:25.571084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:50:25.571094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:50:25.571100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:50:25.573804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.833375 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46c62fbf-0a69-42f6-b25e-85b24cf74ce3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f1058a14b3ef8603e8de8916b37cfb2de17d9855b3a47cbcd447de4d472160\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d2a426a14becafd4034ecb6cda51340f244a88d51949b868baf7a4a27ef5f04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://603b145db89906e07edd2a93df712e88995af6ddb3b8f8cc52b90192d77e34e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.849289 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.863358 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-t9hvz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"08074f18-fe84-4d7b-8327-9696cbe78f38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:51:19Z\\\",\\\"message\\\":\\\"2025-12-03T06:50:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e\\\\n2025-12-03T06:50:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_86f7698f-1bbc-4d41-84b9-d4ef05799a1e to /host/opt/cni/bin/\\\\n2025-12-03T06:50:34Z [verbose] multus-daemon started\\\\n2025-12-03T06:50:34Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:51:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:51:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d6hn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-t9hvz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.882618 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a3371460-14a2-409b-9b8e-603481a252ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc514bafe0794bdcfd42607f769c7193f61618490f1a855c85f810269d6f20ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://960cc6fb539549bd32fa34628b941490126236a5a977376ad7a0694fdc15d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q58cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxmvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.893175 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82db1367-11c6-477b-80dd-8a113fad77b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd8a88432717d9849f1b27b6a16f0e9b876e87412627c730af11adcb8908e952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e141b005fd91066844de73100bd9a84d165c43f83153c4b319eea91794ebd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.905077 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.905341 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.905424 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.905502 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.905570 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:47Z","lastTransitionTime":"2025-12-03T06:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.905187 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff7acd9-0672-41f4-9064-8cd05e75f2ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb207b86f4f39382342c203fcec8153664fffdb7d0d82f07345fca284233edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4de323d182cc578853dba451509cccd5828fc51890e7b0347d7f785923c550c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1838d29619652dd6b98546f8a82115b8ae7b42965323ace911d03b6f112c835c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e545f3b668f70cb43dc9705d4e549c5e2d9394cefc9c63b326a14a94a1f85186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.916937 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6965cc663c41167b04ae3983fe384cc1393c884ec4870e638416ba5e1c231b51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.936552 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81237850-a445-4887-86e0-23bb0fa052c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047f0da81a4983f53eeeb33cf5c296e3f6af52cc7155f685f4f6517204debf88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:50:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bffe7eba8f36ce1abc48042b89a0065ba8d3e8f035d2b7a0465448f67b395f15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ceecb7358f63baae7aac3f72634ee0a78b20b0255ce5c6f3e9b9578ceb2452\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f1e7e266d384fafb930db9c46448de716a71751390be6df5fa7f7ad0950ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2997ae1a840571233c9b6e279f2766869cea290f80bc81b753891ae8135e7cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62c5678dfa4e10070a0eeac63ab5a34fc2c69011fccc30c1ef71ee1537a57236\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f684f70834ed8fde2b8827437953d0a1448e298d39a5e4e5362252717e052a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:50:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht6gh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-lzmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:47 crc kubenswrapper[4946]: I1203 06:51:47.948228 4946 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e5abff8-4252-4371-803d-d241c81c5910\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:50:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5zt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:50:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hkpdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.009221 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.009302 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.009324 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.009354 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.009376 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.112678 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.112791 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.112821 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.112852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.112878 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.214519 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.214553 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.214561 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.214573 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.214581 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.317834 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.317934 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.317954 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.317978 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.317996 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.421417 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.421489 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.421506 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.421535 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.421552 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.524475 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.524525 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.524539 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.524560 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.524575 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.592622 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:48 crc kubenswrapper[4946]: E1203 06:51:48.592844 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.593147 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:48 crc kubenswrapper[4946]: E1203 06:51:48.593519 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.593269 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:48 crc kubenswrapper[4946]: E1203 06:51:48.594371 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.627090 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.627157 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.627174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.627196 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.627222 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.731441 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.731512 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.731528 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.731559 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.731580 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.834687 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.834725 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.834738 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.834775 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.834788 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.937611 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.937675 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.937698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.937727 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:48 crc kubenswrapper[4946]: I1203 06:51:48.937786 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:48Z","lastTransitionTime":"2025-12-03T06:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.042297 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.042366 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.042385 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.042409 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.042426 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.145054 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.145109 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.145122 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.145140 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.145153 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.248726 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.248852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.248882 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.248913 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.248931 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.353052 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.353135 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.353156 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.353181 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.353199 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.456949 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.457042 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.457070 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.457140 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.457165 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.559982 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.560023 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.560032 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.560047 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.560059 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.592765 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:49 crc kubenswrapper[4946]: E1203 06:51:49.592955 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.662689 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.662761 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.662776 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.662796 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.662809 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.766144 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.766214 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.766236 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.766258 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.766277 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.869974 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.870030 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.870047 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.870072 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.870086 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.972481 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.972549 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.972563 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.972589 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:49 crc kubenswrapper[4946]: I1203 06:51:49.972605 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:49Z","lastTransitionTime":"2025-12-03T06:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.075839 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.075893 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.075902 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.075923 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.075932 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.179642 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.179698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.179729 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.179765 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.179779 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.283548 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.283608 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.283617 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.283630 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.283638 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.386175 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.386215 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.386227 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.386247 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.386261 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.489633 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.489696 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.489712 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.489782 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.489799 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.591772 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.591808 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.591727 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:50 crc kubenswrapper[4946]: E1203 06:51:50.591894 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:50 crc kubenswrapper[4946]: E1203 06:51:50.592042 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:50 crc kubenswrapper[4946]: E1203 06:51:50.592086 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.592730 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.592827 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.592846 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.592903 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.592922 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.695960 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.696017 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.696033 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.696057 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.696074 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.799379 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.799460 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.799481 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.799511 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.799533 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.902955 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.903017 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.903034 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.903060 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:50 crc kubenswrapper[4946]: I1203 06:51:50.903078 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:50Z","lastTransitionTime":"2025-12-03T06:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.005363 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.005425 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.005451 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.005492 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.005511 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.108914 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.108986 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.109006 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.109029 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.109045 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.212494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.212549 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.212562 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.212581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.212593 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.224702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.224912 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.225417 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.225514 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.225541 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.245852 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.251536 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.251612 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.251632 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.251657 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.251677 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.273260 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.278433 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.278502 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.278526 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.278556 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.278579 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.295291 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.300798 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.300893 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.300917 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.300948 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.300969 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.322184 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.327794 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.327848 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.327863 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.327886 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.327900 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.342415 4946 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:51:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"53a21bb3-ddb6-4066-b3b5-69f07da5f7ca\\\",\\\"systemUUID\\\":\\\"e734b8d2-0665-40ac-a46d-0333906fc43c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:51:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.342594 4946 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.344519 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.344557 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.344571 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.344590 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.344602 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.447309 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.447356 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.447367 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.447382 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.447393 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.550177 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.550243 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.550265 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.550291 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.550309 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.592433 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.592627 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.652936 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.652988 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.653002 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.653022 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.653035 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.756015 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.756136 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.756155 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.756179 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.756197 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.791419 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.791702 4946 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:51:51 crc kubenswrapper[4946]: E1203 06:51:51.791838 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs podName:2e5abff8-4252-4371-803d-d241c81c5910 nodeName:}" failed. No retries permitted until 2025-12-03 06:52:55.791812325 +0000 UTC m=+168.588502474 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs") pod "network-metrics-daemon-hkpdf" (UID: "2e5abff8-4252-4371-803d-d241c81c5910") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.859901 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.859961 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.859980 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.860006 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.860024 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.962938 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.963009 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.963031 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.963057 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:51 crc kubenswrapper[4946]: I1203 06:51:51.963079 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:51Z","lastTransitionTime":"2025-12-03T06:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.066085 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.066160 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.066171 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.066214 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.066226 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.169585 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.169667 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.169689 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.169714 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.169731 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.273008 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.273075 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.273114 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.273149 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.273171 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.386893 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.386973 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.386996 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.387025 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.387047 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.490418 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.490506 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.490557 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.490580 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.490599 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.591840 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.591842 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:52 crc kubenswrapper[4946]: E1203 06:51:52.592015 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.592328 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:52 crc kubenswrapper[4946]: E1203 06:51:52.592732 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:52 crc kubenswrapper[4946]: E1203 06:51:52.592971 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.594364 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.594407 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.594417 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.594435 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.594449 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.697367 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.697439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.697457 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.697481 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.697499 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.800209 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.800302 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.800313 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.800328 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.800336 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.903471 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.903537 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.903547 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.903572 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:52 crc kubenswrapper[4946]: I1203 06:51:52.903584 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:52Z","lastTransitionTime":"2025-12-03T06:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.006508 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.006550 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.006560 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.006576 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.006585 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.108847 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.108889 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.108898 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.108913 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.108921 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.216942 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.217016 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.217037 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.217065 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.217087 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.320541 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.320616 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.320640 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.320667 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.320687 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.424186 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.424220 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.424228 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.424241 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.424249 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.526716 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.526779 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.526790 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.526803 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.526812 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.592610 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:53 crc kubenswrapper[4946]: E1203 06:51:53.592807 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.630005 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.630048 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.630058 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.630076 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.630086 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.733582 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.733635 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.733651 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.733674 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.733693 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.837023 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.837079 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.837094 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.837118 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.837134 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.939976 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.940013 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.940022 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.940036 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:53 crc kubenswrapper[4946]: I1203 06:51:53.940045 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:53Z","lastTransitionTime":"2025-12-03T06:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.042698 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.042818 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.042842 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.042870 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.042890 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.145001 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.145038 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.145046 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.145061 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.145072 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.248598 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.248676 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.248690 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.248711 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.248729 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.353059 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.353174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.353194 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.353222 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.353244 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.456381 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.456438 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.456450 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.456467 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.456479 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.559365 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.559429 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.559464 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.559493 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.559513 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.592226 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.592284 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.592242 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:54 crc kubenswrapper[4946]: E1203 06:51:54.592375 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:54 crc kubenswrapper[4946]: E1203 06:51:54.592492 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:54 crc kubenswrapper[4946]: E1203 06:51:54.592585 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.662103 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.662183 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.662205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.662243 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.662268 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.765227 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.765279 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.765289 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.765306 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.765316 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.869205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.869254 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.869269 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.869291 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.869308 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.972217 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.972284 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.972303 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.972334 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:54 crc kubenswrapper[4946]: I1203 06:51:54.972353 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:54Z","lastTransitionTime":"2025-12-03T06:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.076813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.076876 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.076893 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.076917 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.076939 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.179793 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.179939 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.179958 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.179985 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.180002 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.283478 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.283542 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.283566 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.283666 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.283698 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.387460 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.387529 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.387555 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.387608 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.387637 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.490960 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.491049 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.491075 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.491108 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.491165 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.591995 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:55 crc kubenswrapper[4946]: E1203 06:51:55.592173 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.593348 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.593403 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.593420 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.593443 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.593459 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.696493 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.696545 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.696554 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.696568 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.696578 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.799846 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.799886 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.799899 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.799917 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.799928 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.903591 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.903652 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.903687 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.903716 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:55 crc kubenswrapper[4946]: I1203 06:51:55.903767 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:55Z","lastTransitionTime":"2025-12-03T06:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.006333 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.006390 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.006402 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.006420 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.006432 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.109520 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.109570 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.109585 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.109605 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.109617 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.212462 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.212530 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.212547 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.212571 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.212589 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.315333 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.315388 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.315405 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.315430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.315447 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.418930 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.418985 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.419006 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.419055 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.419083 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.522430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.522510 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.522534 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.522568 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.522592 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.592585 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.592681 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.592710 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:56 crc kubenswrapper[4946]: E1203 06:51:56.593132 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:56 crc kubenswrapper[4946]: E1203 06:51:56.593296 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:56 crc kubenswrapper[4946]: E1203 06:51:56.593421 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.625659 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.625702 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.625710 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.625730 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.625866 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.729487 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.729543 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.729559 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.729576 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.729588 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.833452 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.833530 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.833556 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.833583 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.833605 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.936458 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.936517 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.936540 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.936568 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:56 crc kubenswrapper[4946]: I1203 06:51:56.936588 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:56Z","lastTransitionTime":"2025-12-03T06:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.039813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.039913 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.039935 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.039992 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.040011 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.142713 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.142827 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.142852 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.142883 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.142906 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.246519 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.246577 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.246590 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.246607 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.246617 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.349399 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.349445 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.349460 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.349482 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.349497 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.453554 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.453612 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.453629 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.453654 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.453671 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.555790 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.555836 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.555849 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.555865 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.555877 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.592174 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:57 crc kubenswrapper[4946]: E1203 06:51:57.592609 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.593376 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 06:51:57 crc kubenswrapper[4946]: E1203 06:51:57.593769 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6fppr_openshift-ovn-kubernetes(58ed82b9-82e5-4d9a-b331-80c32397dc43)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.637989 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podStartSLOduration=85.637972918 podStartE2EDuration="1m25.637972918s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.637855174 +0000 UTC m=+110.434545313" watchObservedRunningTime="2025-12-03 06:51:57.637972918 +0000 UTC m=+110.434663017" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.647428 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-8plsb" podStartSLOduration=85.64740513 podStartE2EDuration="1m25.64740513s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.647258617 +0000 UTC m=+110.443948726" watchObservedRunningTime="2025-12-03 06:51:57.64740513 +0000 UTC m=+110.444095280" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.659662 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.659822 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.659847 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.659933 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.660010 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.675390 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-2b4cb" podStartSLOduration=85.675362651 podStartE2EDuration="1m25.675362651s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.675138685 +0000 UTC m=+110.471828884" watchObservedRunningTime="2025-12-03 06:51:57.675362651 +0000 UTC m=+110.472052800" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.752579 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=89.752556032 podStartE2EDuration="1m29.752556032s" podCreationTimestamp="2025-12-03 06:50:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.752547571 +0000 UTC m=+110.549237740" watchObservedRunningTime="2025-12-03 06:51:57.752556032 +0000 UTC m=+110.549246181" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.764273 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.764321 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.764334 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.764353 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.764366 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.810648 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-t9hvz" podStartSLOduration=85.81062686 podStartE2EDuration="1m25.81062686s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.798881704 +0000 UTC m=+110.595571824" watchObservedRunningTime="2025-12-03 06:51:57.81062686 +0000 UTC m=+110.607316969" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.839729 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.83970636 podStartE2EDuration="1m32.83970636s" podCreationTimestamp="2025-12-03 06:50:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.839156995 +0000 UTC m=+110.635847114" watchObservedRunningTime="2025-12-03 06:51:57.83970636 +0000 UTC m=+110.636396489" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.839987 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxmvh" podStartSLOduration=84.839978807 podStartE2EDuration="1m24.839978807s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.810324271 +0000 UTC m=+110.607014380" watchObservedRunningTime="2025-12-03 06:51:57.839978807 +0000 UTC m=+110.636668926" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.866720 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.868173 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.868205 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.868229 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.868243 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.882263 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.882247401 podStartE2EDuration="1m28.882247401s" podCreationTimestamp="2025-12-03 06:50:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.86283713 +0000 UTC m=+110.659527239" watchObservedRunningTime="2025-12-03 06:51:57.882247401 +0000 UTC m=+110.678937510" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.899497 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-lzmlt" podStartSLOduration=85.899479243 podStartE2EDuration="1m25.899479243s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.899421602 +0000 UTC m=+110.696111721" watchObservedRunningTime="2025-12-03 06:51:57.899479243 +0000 UTC m=+110.696169352" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.933014 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=42.932995333 podStartE2EDuration="42.932995333s" podCreationTimestamp="2025-12-03 06:51:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.920868077 +0000 UTC m=+110.717558196" watchObservedRunningTime="2025-12-03 06:51:57.932995333 +0000 UTC m=+110.729685452" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.933389 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=62.933385603 podStartE2EDuration="1m2.933385603s" podCreationTimestamp="2025-12-03 06:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:51:57.933082295 +0000 UTC m=+110.729772404" watchObservedRunningTime="2025-12-03 06:51:57.933385603 +0000 UTC m=+110.730075712" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.970870 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.970960 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.970975 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.970991 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:57 crc kubenswrapper[4946]: I1203 06:51:57.971002 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:57Z","lastTransitionTime":"2025-12-03T06:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.074330 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.074439 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.074463 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.074494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.074517 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.177430 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.177468 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.177475 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.177491 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.177500 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.280538 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.280577 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.280585 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.280600 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.280608 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.383501 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.383546 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.383561 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.383581 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.383595 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.486732 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.486830 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.486853 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.486876 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.486893 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.592540 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.592628 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.592843 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:51:58 crc kubenswrapper[4946]: E1203 06:51:58.593063 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:51:58 crc kubenswrapper[4946]: E1203 06:51:58.593180 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:51:58 crc kubenswrapper[4946]: E1203 06:51:58.593328 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.598363 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.598417 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.598433 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.598454 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.598471 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.706962 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.707018 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.707035 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.707064 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.707080 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.810507 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.810577 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.810596 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.810621 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.810640 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.913596 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.913680 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.913701 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.913730 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:58 crc kubenswrapper[4946]: I1203 06:51:58.913805 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:58Z","lastTransitionTime":"2025-12-03T06:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.018115 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.018174 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.018192 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.018215 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.018233 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.121884 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.121972 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.122020 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.122052 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.122075 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.225198 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.225244 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.225257 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.225273 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.225285 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.329392 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.329449 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.329473 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.329502 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.329524 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.431714 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.431820 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.431841 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.431867 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.431885 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.536903 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.537008 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.537034 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.537064 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.537086 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.592250 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:51:59 crc kubenswrapper[4946]: E1203 06:51:59.592431 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.639501 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.639564 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.639582 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.639609 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.639627 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.742626 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.742694 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.742717 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.742825 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.742855 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.845563 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.845610 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.845626 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.845647 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.845664 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.948589 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.948655 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.948672 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.948699 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:51:59 crc kubenswrapper[4946]: I1203 06:51:59.948716 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:51:59Z","lastTransitionTime":"2025-12-03T06:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.051624 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.051695 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.051718 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.051794 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.051831 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.154669 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.154770 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.154792 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.154823 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.154845 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.258236 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.258304 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.258329 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.258357 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.258381 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.361423 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.361517 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.361541 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.361571 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.361593 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.464671 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.465163 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.465226 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.465259 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.465282 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.567998 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.568070 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.568096 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.568126 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.568147 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.592617 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.592821 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.592667 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:00 crc kubenswrapper[4946]: E1203 06:52:00.593061 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:00 crc kubenswrapper[4946]: E1203 06:52:00.593179 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:00 crc kubenswrapper[4946]: E1203 06:52:00.592946 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.671501 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.671541 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.671553 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.671570 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.671582 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.775383 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.775450 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.775470 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.775495 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.775512 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.879298 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.879375 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.879400 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.879432 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.879468 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.983108 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.983153 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.983164 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.983182 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:00 crc kubenswrapper[4946]: I1203 06:52:00.983193 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:00Z","lastTransitionTime":"2025-12-03T06:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.085832 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.085894 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.085913 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.085937 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.085958 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:01Z","lastTransitionTime":"2025-12-03T06:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.188813 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.188862 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.188877 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.188898 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.189088 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:01Z","lastTransitionTime":"2025-12-03T06:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.291798 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.291861 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.291884 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.291913 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.291936 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:01Z","lastTransitionTime":"2025-12-03T06:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.395179 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.395234 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.395251 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.395276 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.395293 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:01Z","lastTransitionTime":"2025-12-03T06:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.498416 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.498477 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.498494 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.498516 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.498531 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:01Z","lastTransitionTime":"2025-12-03T06:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.591882 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:01 crc kubenswrapper[4946]: E1203 06:52:01.592418 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.601863 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.601890 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.601899 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.601912 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.601922 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:01Z","lastTransitionTime":"2025-12-03T06:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.695984 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.696033 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.696051 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.696073 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.696089 4946 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:52:01Z","lastTransitionTime":"2025-12-03T06:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.754025 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq"] Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.754521 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.756831 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.757017 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.757211 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.757415 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.803124 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.803209 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.803326 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.803533 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-service-ca\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.803585 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.905023 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-service-ca\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.905094 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.905151 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.905221 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.905292 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.905313 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.905915 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.906427 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-service-ca\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.916120 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:01 crc kubenswrapper[4946]: I1203 06:52:01.934827 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5132c71b-9dd5-4b3c-b9bd-f951cb0b6548-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-smssq\" (UID: \"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:02 crc kubenswrapper[4946]: I1203 06:52:02.078996 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" Dec 03 06:52:02 crc kubenswrapper[4946]: W1203 06:52:02.105007 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5132c71b_9dd5_4b3c_b9bd_f951cb0b6548.slice/crio-cf092082ab1cfd2ddf7f96ceb83ef121fb4bacde620e1ca5fe683dc8e21218ec WatchSource:0}: Error finding container cf092082ab1cfd2ddf7f96ceb83ef121fb4bacde620e1ca5fe683dc8e21218ec: Status 404 returned error can't find the container with id cf092082ab1cfd2ddf7f96ceb83ef121fb4bacde620e1ca5fe683dc8e21218ec Dec 03 06:52:02 crc kubenswrapper[4946]: I1203 06:52:02.592491 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:02 crc kubenswrapper[4946]: I1203 06:52:02.592497 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:02 crc kubenswrapper[4946]: I1203 06:52:02.592660 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:02 crc kubenswrapper[4946]: E1203 06:52:02.592916 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:02 crc kubenswrapper[4946]: E1203 06:52:02.593232 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:02 crc kubenswrapper[4946]: E1203 06:52:02.593458 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:02 crc kubenswrapper[4946]: I1203 06:52:02.988867 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" event={"ID":"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548","Type":"ContainerStarted","Data":"83629de3dac68df3892b9c809c0c49dce8b4ef2c2beaa34ac5d4ce94c4673f7c"} Dec 03 06:52:02 crc kubenswrapper[4946]: I1203 06:52:02.988940 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" event={"ID":"5132c71b-9dd5-4b3c-b9bd-f951cb0b6548","Type":"ContainerStarted","Data":"cf092082ab1cfd2ddf7f96ceb83ef121fb4bacde620e1ca5fe683dc8e21218ec"} Dec 03 06:52:03 crc kubenswrapper[4946]: I1203 06:52:03.009970 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-smssq" podStartSLOduration=91.00994393 podStartE2EDuration="1m31.00994393s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:03.008396028 +0000 UTC m=+115.805086188" watchObservedRunningTime="2025-12-03 06:52:03.00994393 +0000 UTC m=+115.806634069" Dec 03 06:52:03 crc kubenswrapper[4946]: I1203 06:52:03.592080 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:03 crc kubenswrapper[4946]: E1203 06:52:03.592253 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:04 crc kubenswrapper[4946]: I1203 06:52:04.591892 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:04 crc kubenswrapper[4946]: I1203 06:52:04.591985 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:04 crc kubenswrapper[4946]: I1203 06:52:04.591908 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:04 crc kubenswrapper[4946]: E1203 06:52:04.592225 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:04 crc kubenswrapper[4946]: E1203 06:52:04.592316 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:04 crc kubenswrapper[4946]: E1203 06:52:04.592125 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:05 crc kubenswrapper[4946]: I1203 06:52:05.592493 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:05 crc kubenswrapper[4946]: E1203 06:52:05.592676 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:06 crc kubenswrapper[4946]: I1203 06:52:06.591882 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:06 crc kubenswrapper[4946]: I1203 06:52:06.592025 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:06 crc kubenswrapper[4946]: I1203 06:52:06.592153 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:06 crc kubenswrapper[4946]: E1203 06:52:06.592045 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:06 crc kubenswrapper[4946]: E1203 06:52:06.592333 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:06 crc kubenswrapper[4946]: E1203 06:52:06.592460 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:07 crc kubenswrapper[4946]: I1203 06:52:07.006142 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/1.log" Dec 03 06:52:07 crc kubenswrapper[4946]: I1203 06:52:07.006779 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/0.log" Dec 03 06:52:07 crc kubenswrapper[4946]: I1203 06:52:07.006885 4946 generic.go:334] "Generic (PLEG): container finished" podID="08074f18-fe84-4d7b-8327-9696cbe78f38" containerID="e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59" exitCode=1 Dec 03 06:52:07 crc kubenswrapper[4946]: I1203 06:52:07.006932 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t9hvz" event={"ID":"08074f18-fe84-4d7b-8327-9696cbe78f38","Type":"ContainerDied","Data":"e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59"} Dec 03 06:52:07 crc kubenswrapper[4946]: I1203 06:52:07.007021 4946 scope.go:117] "RemoveContainer" containerID="79a5a6f0ae3bab3bbadda6fdf5fa14d1dd9adf798e6e4e5dcf2382cdf357a7f7" Dec 03 06:52:07 crc kubenswrapper[4946]: I1203 06:52:07.007926 4946 scope.go:117] "RemoveContainer" containerID="e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59" Dec 03 06:52:07 crc kubenswrapper[4946]: E1203 06:52:07.008267 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-t9hvz_openshift-multus(08074f18-fe84-4d7b-8327-9696cbe78f38)\"" pod="openshift-multus/multus-t9hvz" podUID="08074f18-fe84-4d7b-8327-9696cbe78f38" Dec 03 06:52:07 crc kubenswrapper[4946]: I1203 06:52:07.593124 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:07 crc kubenswrapper[4946]: E1203 06:52:07.594280 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:07 crc kubenswrapper[4946]: E1203 06:52:07.613121 4946 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 06:52:07 crc kubenswrapper[4946]: E1203 06:52:07.699567 4946 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:52:08 crc kubenswrapper[4946]: I1203 06:52:08.013506 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/1.log" Dec 03 06:52:08 crc kubenswrapper[4946]: I1203 06:52:08.592398 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:08 crc kubenswrapper[4946]: I1203 06:52:08.592541 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:08 crc kubenswrapper[4946]: E1203 06:52:08.592624 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:08 crc kubenswrapper[4946]: I1203 06:52:08.592700 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:08 crc kubenswrapper[4946]: E1203 06:52:08.593212 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:08 crc kubenswrapper[4946]: E1203 06:52:08.593368 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:09 crc kubenswrapper[4946]: I1203 06:52:09.592917 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:09 crc kubenswrapper[4946]: E1203 06:52:09.593298 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:10 crc kubenswrapper[4946]: I1203 06:52:10.592578 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:10 crc kubenswrapper[4946]: E1203 06:52:10.593057 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:10 crc kubenswrapper[4946]: I1203 06:52:10.592728 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:10 crc kubenswrapper[4946]: E1203 06:52:10.593163 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:10 crc kubenswrapper[4946]: I1203 06:52:10.592819 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:10 crc kubenswrapper[4946]: E1203 06:52:10.593220 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:11 crc kubenswrapper[4946]: I1203 06:52:11.592501 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:11 crc kubenswrapper[4946]: E1203 06:52:11.592697 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:12 crc kubenswrapper[4946]: I1203 06:52:12.592445 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:12 crc kubenswrapper[4946]: I1203 06:52:12.592459 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:12 crc kubenswrapper[4946]: I1203 06:52:12.592864 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:12 crc kubenswrapper[4946]: E1203 06:52:12.593132 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:12 crc kubenswrapper[4946]: E1203 06:52:12.593204 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:12 crc kubenswrapper[4946]: E1203 06:52:12.593278 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:12 crc kubenswrapper[4946]: I1203 06:52:12.593457 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 06:52:12 crc kubenswrapper[4946]: E1203 06:52:12.701333 4946 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:52:13 crc kubenswrapper[4946]: I1203 06:52:13.031146 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/3.log" Dec 03 06:52:13 crc kubenswrapper[4946]: I1203 06:52:13.033955 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerStarted","Data":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} Dec 03 06:52:13 crc kubenswrapper[4946]: I1203 06:52:13.034442 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:52:13 crc kubenswrapper[4946]: I1203 06:52:13.071019 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podStartSLOduration=101.071001533 podStartE2EDuration="1m41.071001533s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:13.06268411 +0000 UTC m=+125.859374249" watchObservedRunningTime="2025-12-03 06:52:13.071001533 +0000 UTC m=+125.867691642" Dec 03 06:52:13 crc kubenswrapper[4946]: I1203 06:52:13.394846 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hkpdf"] Dec 03 06:52:13 crc kubenswrapper[4946]: I1203 06:52:13.395048 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:13 crc kubenswrapper[4946]: E1203 06:52:13.395179 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:13 crc kubenswrapper[4946]: I1203 06:52:13.592674 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:13 crc kubenswrapper[4946]: E1203 06:52:13.592989 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:14 crc kubenswrapper[4946]: I1203 06:52:14.592193 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:14 crc kubenswrapper[4946]: I1203 06:52:14.592267 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:14 crc kubenswrapper[4946]: E1203 06:52:14.592351 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:14 crc kubenswrapper[4946]: E1203 06:52:14.592439 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:15 crc kubenswrapper[4946]: I1203 06:52:15.591989 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:15 crc kubenswrapper[4946]: I1203 06:52:15.592116 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:15 crc kubenswrapper[4946]: E1203 06:52:15.592188 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:15 crc kubenswrapper[4946]: E1203 06:52:15.592403 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:16 crc kubenswrapper[4946]: I1203 06:52:16.592134 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:16 crc kubenswrapper[4946]: E1203 06:52:16.592275 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:16 crc kubenswrapper[4946]: I1203 06:52:16.592137 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:16 crc kubenswrapper[4946]: E1203 06:52:16.592425 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:17 crc kubenswrapper[4946]: I1203 06:52:17.592326 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:17 crc kubenswrapper[4946]: E1203 06:52:17.594499 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:17 crc kubenswrapper[4946]: I1203 06:52:17.594660 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:17 crc kubenswrapper[4946]: E1203 06:52:17.594895 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:17 crc kubenswrapper[4946]: E1203 06:52:17.702007 4946 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:52:18 crc kubenswrapper[4946]: I1203 06:52:18.592391 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:18 crc kubenswrapper[4946]: I1203 06:52:18.592415 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:18 crc kubenswrapper[4946]: E1203 06:52:18.592561 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:18 crc kubenswrapper[4946]: E1203 06:52:18.592923 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:19 crc kubenswrapper[4946]: I1203 06:52:19.592330 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:19 crc kubenswrapper[4946]: E1203 06:52:19.592559 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:19 crc kubenswrapper[4946]: I1203 06:52:19.592963 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:19 crc kubenswrapper[4946]: E1203 06:52:19.593116 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:20 crc kubenswrapper[4946]: I1203 06:52:20.591948 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:20 crc kubenswrapper[4946]: E1203 06:52:20.592358 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:20 crc kubenswrapper[4946]: I1203 06:52:20.592875 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:20 crc kubenswrapper[4946]: I1203 06:52:20.592997 4946 scope.go:117] "RemoveContainer" containerID="e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59" Dec 03 06:52:20 crc kubenswrapper[4946]: E1203 06:52:20.593073 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:21 crc kubenswrapper[4946]: I1203 06:52:21.176781 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/1.log" Dec 03 06:52:21 crc kubenswrapper[4946]: I1203 06:52:21.176857 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t9hvz" event={"ID":"08074f18-fe84-4d7b-8327-9696cbe78f38","Type":"ContainerStarted","Data":"7570afed4df875e4bd9bb2d80ac89f61f8a0151c58f9912a40f360f2d41528d2"} Dec 03 06:52:21 crc kubenswrapper[4946]: I1203 06:52:21.591961 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:21 crc kubenswrapper[4946]: E1203 06:52:21.592703 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:21 crc kubenswrapper[4946]: I1203 06:52:21.591962 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:21 crc kubenswrapper[4946]: E1203 06:52:21.593001 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:22 crc kubenswrapper[4946]: I1203 06:52:22.592497 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:22 crc kubenswrapper[4946]: I1203 06:52:22.592563 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:22 crc kubenswrapper[4946]: E1203 06:52:22.592723 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:22 crc kubenswrapper[4946]: E1203 06:52:22.592845 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:22 crc kubenswrapper[4946]: E1203 06:52:22.703337 4946 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:52:23 crc kubenswrapper[4946]: I1203 06:52:23.592080 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:23 crc kubenswrapper[4946]: I1203 06:52:23.592194 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:23 crc kubenswrapper[4946]: E1203 06:52:23.592309 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:23 crc kubenswrapper[4946]: E1203 06:52:23.592410 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:24 crc kubenswrapper[4946]: I1203 06:52:24.592317 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:24 crc kubenswrapper[4946]: I1203 06:52:24.592386 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:24 crc kubenswrapper[4946]: E1203 06:52:24.592492 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:24 crc kubenswrapper[4946]: E1203 06:52:24.592808 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:25 crc kubenswrapper[4946]: I1203 06:52:25.592191 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:25 crc kubenswrapper[4946]: I1203 06:52:25.592341 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:25 crc kubenswrapper[4946]: E1203 06:52:25.592468 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:25 crc kubenswrapper[4946]: E1203 06:52:25.592722 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:26 crc kubenswrapper[4946]: I1203 06:52:26.592706 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:26 crc kubenswrapper[4946]: I1203 06:52:26.592707 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:26 crc kubenswrapper[4946]: E1203 06:52:26.592959 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:52:26 crc kubenswrapper[4946]: E1203 06:52:26.593173 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:52:27 crc kubenswrapper[4946]: I1203 06:52:27.591941 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:27 crc kubenswrapper[4946]: I1203 06:52:27.591963 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:27 crc kubenswrapper[4946]: E1203 06:52:27.594603 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hkpdf" podUID="2e5abff8-4252-4371-803d-d241c81c5910" Dec 03 06:52:27 crc kubenswrapper[4946]: E1203 06:52:27.594787 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:52:28 crc kubenswrapper[4946]: I1203 06:52:28.592680 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:28 crc kubenswrapper[4946]: I1203 06:52:28.592777 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:28 crc kubenswrapper[4946]: I1203 06:52:28.596206 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 06:52:28 crc kubenswrapper[4946]: I1203 06:52:28.596359 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 06:52:28 crc kubenswrapper[4946]: I1203 06:52:28.596531 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 06:52:28 crc kubenswrapper[4946]: I1203 06:52:28.596599 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 06:52:29 crc kubenswrapper[4946]: I1203 06:52:29.592614 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:29 crc kubenswrapper[4946]: I1203 06:52:29.592615 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:29 crc kubenswrapper[4946]: I1203 06:52:29.594926 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 06:52:29 crc kubenswrapper[4946]: I1203 06:52:29.595216 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.506688 4946 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.563256 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vcrrg"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.563991 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.573457 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.573767 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.575233 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.576021 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.581728 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.581884 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.582251 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.582580 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.585823 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.586306 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f588s"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.586845 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.587244 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.587405 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.587893 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.602267 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.603026 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.603361 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.603520 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7nc8s"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.604200 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.604565 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.604935 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.605634 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.605851 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.605873 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.605996 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.606087 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.606107 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.606215 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.606256 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.606466 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.606727 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.605644 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.614236 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.608888 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.615566 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vcrrg"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.615600 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ncr4w"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.615922 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.616511 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f59l8"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.616890 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-zkld2"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.617183 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w6x6n"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.617472 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-s7slv"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.617848 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.618005 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.618235 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ncr4w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.618458 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.614321 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.614414 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.614422 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.614500 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.614541 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.614691 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.614894 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.620043 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.620050 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.620319 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.624758 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-82rd5"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.625269 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.627673 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.627879 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.628055 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.628205 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.659302 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.663356 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.663762 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.663915 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.664003 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.664387 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.664674 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.678979 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.679279 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.679732 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qwfgf"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680097 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680126 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680326 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ff611ff-6a55-435a-9723-df1b38741952-images\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680364 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c199c97-422e-4a3d-a598-f28a651b377f-auth-proxy-config\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680394 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c199c97-422e-4a3d-a598-f28a651b377f-machine-approver-tls\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680442 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c79v5\" (UniqueName: \"kubernetes.io/projected/6ff611ff-6a55-435a-9723-df1b38741952-kube-api-access-c79v5\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680475 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff611ff-6a55-435a-9723-df1b38741952-config\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680484 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680500 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjpb6\" (UniqueName: \"kubernetes.io/projected/6c199c97-422e-4a3d-a598-f28a651b377f-kube-api-access-zjpb6\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680526 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c199c97-422e-4a3d-a598-f28a651b377f-config\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680569 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ff611ff-6a55-435a-9723-df1b38741952-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680692 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.680923 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.681310 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.681639 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.681812 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.681980 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.682318 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.682439 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.682441 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.682530 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.682601 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.682634 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.682670 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.682775 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.683956 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.685271 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.690664 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.691146 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.693053 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.693418 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nt44m"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.693662 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.694720 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.698141 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.706054 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.706503 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.706712 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.706925 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.707103 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.707348 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.707509 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.707790 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.707815 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.708307 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.708659 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.713857 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.715187 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.715430 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.715594 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.715958 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-42855"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.715985 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.716184 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.716439 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.716758 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.716956 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.717872 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.720084 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.720576 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.720672 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.720966 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.728792 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.729537 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.732236 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.732933 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.733237 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.733508 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.736189 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fsx6z"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.737003 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.737623 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.740322 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.743217 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzrcc"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.744597 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.744703 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.745022 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.745262 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.745695 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-l68r8"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.747537 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.748054 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7nc8s"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.749566 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.750873 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.751020 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.752470 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.752711 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.762910 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.764840 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.764894 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.784683 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ncr4w"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.784763 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.785087 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.785575 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.786468 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.787021 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.787424 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.787729 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.787949 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.788255 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.788564 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797055 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-image-import-ca\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797106 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797131 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a4c7b48-79a1-4a55-9012-a50c96908e7a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797151 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-config\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797170 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-audit-dir\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797191 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84f01e09-8152-4a65-b71a-5456ea557a77-serving-cert\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797207 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece5ff47-76df-4fad-9a59-3dbba27b788d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797233 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hckq6\" (UniqueName: \"kubernetes.io/projected/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-kube-api-access-hckq6\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797253 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-dir\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797273 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-service-ca-bundle\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797291 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-serving-cert\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797310 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-serving-cert\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797335 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ff611ff-6a55-435a-9723-df1b38741952-images\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797360 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c199c97-422e-4a3d-a598-f28a651b377f-machine-approver-tls\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797383 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1bd7b04-110e-45fb-b972-c662a2e7c791-audit-dir\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797403 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkcxz\" (UniqueName: \"kubernetes.io/projected/f1bd7b04-110e-45fb-b972-c662a2e7c791-kube-api-access-tkcxz\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797423 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-etcd-client\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797445 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjm4q\" (UniqueName: \"kubernetes.io/projected/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-kube-api-access-jjm4q\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797469 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-config\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797493 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-config\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797525 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-trusted-ca\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797562 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797578 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f27b8b97-2934-4d16-a756-8d402ec86bad-trusted-ca\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797598 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797615 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-audit-policies\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797634 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-etcd-client\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797657 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797693 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f27b8b97-2934-4d16-a756-8d402ec86bad-bound-sa-token\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797710 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gwdg\" (UniqueName: \"kubernetes.io/projected/cf1e2d1c-6794-48f0-9dde-8eacb6a16525-kube-api-access-6gwdg\") pod \"cluster-samples-operator-665b6dd947-rcb2j\" (UID: \"cf1e2d1c-6794-48f0-9dde-8eacb6a16525\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797732 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-serving-cert\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797773 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-oauth-config\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797796 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjpb6\" (UniqueName: \"kubernetes.io/projected/6c199c97-422e-4a3d-a598-f28a651b377f-kube-api-access-zjpb6\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797815 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797836 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjtdz\" (UniqueName: \"kubernetes.io/projected/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-kube-api-access-gjtdz\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797853 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-client-ca\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797868 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-encryption-config\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797887 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-trusted-ca-bundle\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797917 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c199c97-422e-4a3d-a598-f28a651b377f-config\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797937 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9499\" (UniqueName: \"kubernetes.io/projected/ece5ff47-76df-4fad-9a59-3dbba27b788d-kube-api-access-c9499\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797962 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a4c7b48-79a1-4a55-9012-a50c96908e7a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.797988 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798004 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798020 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798056 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-policies\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798073 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798091 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/84f01e09-8152-4a65-b71a-5456ea557a77-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798116 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqckd\" (UniqueName: \"kubernetes.io/projected/4a4c7b48-79a1-4a55-9012-a50c96908e7a-kube-api-access-xqckd\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798134 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798157 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ff611ff-6a55-435a-9723-df1b38741952-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798172 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece5ff47-76df-4fad-9a59-3dbba27b788d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798215 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qms6c\" (UniqueName: \"kubernetes.io/projected/84f01e09-8152-4a65-b71a-5456ea557a77-kube-api-access-qms6c\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798232 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f27b8b97-2934-4d16-a756-8d402ec86bad-metrics-tls\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798333 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c199c97-422e-4a3d-a598-f28a651b377f-auth-proxy-config\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798359 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798469 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gssj8\" (UniqueName: \"kubernetes.io/projected/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-kube-api-access-gssj8\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798494 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798591 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-serving-cert\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798609 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798627 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-config\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.803478 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f588s"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.798700 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-node-pullsecrets\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.823971 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.824202 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.824779 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-audit\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.825292 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.825410 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.825584 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.830021 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.834059 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.834400 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c199c97-422e-4a3d-a598-f28a651b377f-config\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.834984 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.835336 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840266 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840313 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-encryption-config\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840564 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-etcd-serving-ca\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840649 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c79v5\" (UniqueName: \"kubernetes.io/projected/6ff611ff-6a55-435a-9723-df1b38741952-kube-api-access-c79v5\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840697 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blgxs\" (UniqueName: \"kubernetes.io/projected/f2d67105-8c6c-44b1-9baa-71424ee41402-kube-api-access-blgxs\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840722 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25c62\" (UniqueName: \"kubernetes.io/projected/f27b8b97-2934-4d16-a756-8d402ec86bad-kube-api-access-25c62\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840791 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c970075-0d42-45b3-a56f-f0686e7a2455-serving-cert\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840814 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6nk5\" (UniqueName: \"kubernetes.io/projected/ca8d6db9-e160-4ad7-a399-3dc1618add0f-kube-api-access-m6nk5\") pod \"downloads-7954f5f757-ncr4w\" (UID: \"ca8d6db9-e160-4ad7-a399-3dc1618add0f\") " pod="openshift-console/downloads-7954f5f757-ncr4w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840848 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff611ff-6a55-435a-9723-df1b38741952-config\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.840867 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-config\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.841010 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ff611ff-6a55-435a-9723-df1b38741952-images\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.841471 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-client-ca\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.841500 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.841500 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-serving-cert\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.841683 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.841628 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff611ff-6a55-435a-9723-df1b38741952-config\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.841798 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtlvs\" (UniqueName: \"kubernetes.io/projected/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-kube-api-access-qtlvs\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.841951 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.842116 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6jpb\" (UniqueName: \"kubernetes.io/projected/6c970075-0d42-45b3-a56f-f0686e7a2455-kube-api-access-d6jpb\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.842156 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-s7slv"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.842169 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-serving-cert\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.842197 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-config\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.842285 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf1e2d1c-6794-48f0-9dde-8eacb6a16525-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rcb2j\" (UID: \"cf1e2d1c-6794-48f0-9dde-8eacb6a16525\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.842317 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-service-ca\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.842364 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-oauth-serving-cert\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.844138 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.844566 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zkld2"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.844671 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c199c97-422e-4a3d-a598-f28a651b377f-auth-proxy-config\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.844953 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c199c97-422e-4a3d-a598-f28a651b377f-machine-approver-tls\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.845926 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.849703 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.851995 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.849207 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.854619 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.855187 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.855211 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ff611ff-6a55-435a-9723-df1b38741952-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.855935 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nt44m"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.858206 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzrcc"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.859978 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f59l8"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.861025 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.861929 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7sng8"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.862675 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w6x6n"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.862873 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.863644 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.864409 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.864587 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.865585 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.866051 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.866140 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.866684 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.867269 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.867581 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.868363 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.869256 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.869664 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.871152 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.871313 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.879004 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.879474 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.890186 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.890253 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-82rd5"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.890269 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qsg2m"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.891309 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.891326 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.891340 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8rwkt"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.892252 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.892615 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.892830 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qsg2m" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.893942 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.898913 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fsx6z"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.900232 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.909426 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7sng8"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.917234 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.917537 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.918763 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.919803 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.920780 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qwfgf"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.921840 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.922907 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.923919 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-55ddd"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.925384 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qsg2m"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.925487 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.926073 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.926180 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.927368 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.928880 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.930010 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.931133 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-55ddd"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.932513 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.933597 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-42855"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.934782 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-2q2qg"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.935769 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-2q2qg"] Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.935781 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944180 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-serving-cert\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944206 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944225 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-config\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944241 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-node-pullsecrets\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944259 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-audit\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944286 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944309 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a3aa4249-2022-45e7-844a-25e6b14f0661-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944328 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944343 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-encryption-config\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944359 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8afa818f-4048-49d9-a54a-3ba48f3de8cd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944378 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-etcd-serving-ca\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944392 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8afa818f-4048-49d9-a54a-3ba48f3de8cd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944420 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blgxs\" (UniqueName: \"kubernetes.io/projected/f2d67105-8c6c-44b1-9baa-71424ee41402-kube-api-access-blgxs\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944437 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25c62\" (UniqueName: \"kubernetes.io/projected/f27b8b97-2934-4d16-a756-8d402ec86bad-kube-api-access-25c62\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944451 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c970075-0d42-45b3-a56f-f0686e7a2455-serving-cert\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944469 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6nk5\" (UniqueName: \"kubernetes.io/projected/ca8d6db9-e160-4ad7-a399-3dc1618add0f-kube-api-access-m6nk5\") pod \"downloads-7954f5f757-ncr4w\" (UID: \"ca8d6db9-e160-4ad7-a399-3dc1618add0f\") " pod="openshift-console/downloads-7954f5f757-ncr4w" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944484 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-metrics-certs\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944500 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-config\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944518 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t59mq\" (UniqueName: \"kubernetes.io/projected/7ef124c8-9694-4320-bdda-2ecc10489225-kube-api-access-t59mq\") pod \"dns-operator-744455d44c-nt44m\" (UID: \"7ef124c8-9694-4320-bdda-2ecc10489225\") " pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944540 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-serving-cert\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944554 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944570 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/670f4f60-09f3-4a59-bc86-b34fe9466e2e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944586 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-client-ca\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944602 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtlvs\" (UniqueName: \"kubernetes.io/projected/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-kube-api-access-qtlvs\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944619 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkfpf\" (UniqueName: \"kubernetes.io/projected/a3aa4249-2022-45e7-844a-25e6b14f0661-kube-api-access-mkfpf\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944635 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944651 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6jpb\" (UniqueName: \"kubernetes.io/projected/6c970075-0d42-45b3-a56f-f0686e7a2455-kube-api-access-d6jpb\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944665 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8afa818f-4048-49d9-a54a-3ba48f3de8cd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944683 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-config\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944700 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf1e2d1c-6794-48f0-9dde-8eacb6a16525-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rcb2j\" (UID: \"cf1e2d1c-6794-48f0-9dde-8eacb6a16525\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944715 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-service-ca\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944732 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-oauth-serving-cert\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944764 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e87e9694-6803-4745-8a15-4065e6722e08-auth-proxy-config\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944782 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-serving-cert\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944811 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-image-import-ca\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944829 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-service-ca\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944847 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-stats-auth\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944863 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ef124c8-9694-4320-bdda-2ecc10489225-metrics-tls\") pod \"dns-operator-744455d44c-nt44m\" (UID: \"7ef124c8-9694-4320-bdda-2ecc10489225\") " pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944890 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a4c7b48-79a1-4a55-9012-a50c96908e7a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944908 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-config\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944926 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-audit-dir\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944943 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944959 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece5ff47-76df-4fad-9a59-3dbba27b788d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944986 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hckq6\" (UniqueName: \"kubernetes.io/projected/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-kube-api-access-hckq6\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945003 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1b905b7b-c2ff-4cf1-9102-d7124d618ec7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fsx6z\" (UID: \"1b905b7b-c2ff-4cf1-9102-d7124d618ec7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945043 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84f01e09-8152-4a65-b71a-5456ea557a77-serving-cert\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945060 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-service-ca-bundle\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945075 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-serving-cert\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945092 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-serving-cert\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945123 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-dir\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945142 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1bd7b04-110e-45fb-b972-c662a2e7c791-audit-dir\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945164 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkcxz\" (UniqueName: \"kubernetes.io/projected/f1bd7b04-110e-45fb-b972-c662a2e7c791-kube-api-access-tkcxz\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945179 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-etcd-client\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945193 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjm4q\" (UniqueName: \"kubernetes.io/projected/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-kube-api-access-jjm4q\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945210 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l58b7\" (UniqueName: \"kubernetes.io/projected/8afa818f-4048-49d9-a54a-3ba48f3de8cd-kube-api-access-l58b7\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945227 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-config\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945242 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-config\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945266 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-config\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945310 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/670f4f60-09f3-4a59-bc86-b34fe9466e2e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945330 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-default-certificate\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945347 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945363 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f27b8b97-2934-4d16-a756-8d402ec86bad-trusted-ca\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945378 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-trusted-ca\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945395 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945413 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-audit-policies\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945435 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-etcd-client\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945460 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e87e9694-6803-4745-8a15-4065e6722e08-proxy-tls\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945480 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945499 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgtlw\" (UniqueName: \"kubernetes.io/projected/56fe8e2e-1586-4fbd-b68b-f207813a7c35-kube-api-access-rgtlw\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945526 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cafb5284-e531-4979-86aa-ff4c626b71be-serving-cert\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945555 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f27b8b97-2934-4d16-a756-8d402ec86bad-bound-sa-token\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945571 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gwdg\" (UniqueName: \"kubernetes.io/projected/cf1e2d1c-6794-48f0-9dde-8eacb6a16525-kube-api-access-6gwdg\") pod \"cluster-samples-operator-665b6dd947-rcb2j\" (UID: \"cf1e2d1c-6794-48f0-9dde-8eacb6a16525\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945586 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-serving-cert\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945603 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-oauth-config\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945620 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-ca\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945638 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945654 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjtdz\" (UniqueName: \"kubernetes.io/projected/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-kube-api-access-gjtdz\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945668 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-client-ca\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945683 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-encryption-config\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945698 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-trusted-ca-bundle\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945729 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9499\" (UniqueName: \"kubernetes.io/projected/ece5ff47-76df-4fad-9a59-3dbba27b788d-kube-api-access-c9499\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945759 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a4c7b48-79a1-4a55-9012-a50c96908e7a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945777 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdh4p\" (UniqueName: \"kubernetes.io/projected/e87e9694-6803-4745-8a15-4065e6722e08-kube-api-access-fdh4p\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945794 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945818 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945834 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945850 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plgvs\" (UniqueName: \"kubernetes.io/projected/1b905b7b-c2ff-4cf1-9102-d7124d618ec7-kube-api-access-plgvs\") pod \"multus-admission-controller-857f4d67dd-fsx6z\" (UID: \"1b905b7b-c2ff-4cf1-9102-d7124d618ec7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945879 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-policies\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945896 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945916 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/84f01e09-8152-4a65-b71a-5456ea557a77-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945941 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqckd\" (UniqueName: \"kubernetes.io/projected/4a4c7b48-79a1-4a55-9012-a50c96908e7a-kube-api-access-xqckd\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945955 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-config\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.945959 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-client\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946015 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/670f4f60-09f3-4a59-bc86-b34fe9466e2e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946043 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946061 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a3aa4249-2022-45e7-844a-25e6b14f0661-proxy-tls\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946078 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e87e9694-6803-4745-8a15-4065e6722e08-images\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946118 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qms6c\" (UniqueName: \"kubernetes.io/projected/84f01e09-8152-4a65-b71a-5456ea557a77-kube-api-access-qms6c\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946147 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece5ff47-76df-4fad-9a59-3dbba27b788d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946169 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwznh\" (UniqueName: \"kubernetes.io/projected/cafb5284-e531-4979-86aa-ff4c626b71be-kube-api-access-nwznh\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946197 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f27b8b97-2934-4d16-a756-8d402ec86bad-metrics-tls\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946219 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946238 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gssj8\" (UniqueName: \"kubernetes.io/projected/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-kube-api-access-gssj8\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946255 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56fe8e2e-1586-4fbd-b68b-f207813a7c35-service-ca-bundle\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946413 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-node-pullsecrets\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946900 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-audit\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.946970 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.947579 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-config\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.947652 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.947804 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.948491 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.949039 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-serving-cert\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.949908 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.950076 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-oauth-serving-cert\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.950097 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-service-ca\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.950115 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-client-ca\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.950713 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ece5ff47-76df-4fad-9a59-3dbba27b788d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.951100 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf1e2d1c-6794-48f0-9dde-8eacb6a16525-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rcb2j\" (UID: \"cf1e2d1c-6794-48f0-9dde-8eacb6a16525\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.944905 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.951674 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-service-ca-bundle\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.951754 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-audit-dir\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.951901 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-etcd-serving-ca\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.952014 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-config\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.952663 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-image-import-ca\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.952671 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece5ff47-76df-4fad-9a59-3dbba27b788d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.952797 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f27b8b97-2934-4d16-a756-8d402ec86bad-trusted-ca\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.952993 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-encryption-config\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.953356 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.953563 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-dir\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.953616 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1bd7b04-110e-45fb-b972-c662a2e7c791-audit-dir\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.953965 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.954099 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-policies\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.954118 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a4c7b48-79a1-4a55-9012-a50c96908e7a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.954555 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1bd7b04-110e-45fb-b972-c662a2e7c791-audit-policies\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.954899 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.955091 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/84f01e09-8152-4a65-b71a-5456ea557a77-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.955118 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c970075-0d42-45b3-a56f-f0686e7a2455-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.955211 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-config\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.955391 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.955411 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.955726 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-trusted-ca\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.956644 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-config\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.956424 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.957500 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-etcd-client\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.957729 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84f01e09-8152-4a65-b71a-5456ea557a77-serving-cert\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.957846 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.958479 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.958595 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-serving-cert\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.958647 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-config\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.958705 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c970075-0d42-45b3-a56f-f0686e7a2455-serving-cert\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.958851 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-trusted-ca-bundle\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.958994 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-client-ca\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.959310 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a4c7b48-79a1-4a55-9012-a50c96908e7a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.959388 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-serving-cert\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.959638 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.959972 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-oauth-config\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.960203 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-etcd-client\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.960451 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-serving-cert\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.960614 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-serving-cert\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.960702 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f1bd7b04-110e-45fb-b972-c662a2e7c791-encryption-config\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.960913 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f27b8b97-2934-4d16-a756-8d402ec86bad-metrics-tls\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.961468 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-serving-cert\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.963892 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.966006 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 06:52:32 crc kubenswrapper[4946]: I1203 06:52:32.987458 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.005884 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.027461 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.045855 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047005 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkfpf\" (UniqueName: \"kubernetes.io/projected/a3aa4249-2022-45e7-844a-25e6b14f0661-kube-api-access-mkfpf\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047047 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8afa818f-4048-49d9-a54a-3ba48f3de8cd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047075 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e87e9694-6803-4745-8a15-4065e6722e08-auth-proxy-config\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047104 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-service-ca\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047128 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ef124c8-9694-4320-bdda-2ecc10489225-metrics-tls\") pod \"dns-operator-744455d44c-nt44m\" (UID: \"7ef124c8-9694-4320-bdda-2ecc10489225\") " pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047146 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-stats-auth\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047183 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1b905b7b-c2ff-4cf1-9102-d7124d618ec7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fsx6z\" (UID: \"1b905b7b-c2ff-4cf1-9102-d7124d618ec7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047226 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l58b7\" (UniqueName: \"kubernetes.io/projected/8afa818f-4048-49d9-a54a-3ba48f3de8cd-kube-api-access-l58b7\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047252 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-config\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047287 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/670f4f60-09f3-4a59-bc86-b34fe9466e2e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047317 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-default-certificate\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047342 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e87e9694-6803-4745-8a15-4065e6722e08-proxy-tls\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047369 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgtlw\" (UniqueName: \"kubernetes.io/projected/56fe8e2e-1586-4fbd-b68b-f207813a7c35-kube-api-access-rgtlw\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047446 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cafb5284-e531-4979-86aa-ff4c626b71be-serving-cert\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047471 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-ca\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047544 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdh4p\" (UniqueName: \"kubernetes.io/projected/e87e9694-6803-4745-8a15-4065e6722e08-kube-api-access-fdh4p\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047574 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plgvs\" (UniqueName: \"kubernetes.io/projected/1b905b7b-c2ff-4cf1-9102-d7124d618ec7-kube-api-access-plgvs\") pod \"multus-admission-controller-857f4d67dd-fsx6z\" (UID: \"1b905b7b-c2ff-4cf1-9102-d7124d618ec7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047632 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-client\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047657 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/670f4f60-09f3-4a59-bc86-b34fe9466e2e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047689 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a3aa4249-2022-45e7-844a-25e6b14f0661-proxy-tls\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047764 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e87e9694-6803-4745-8a15-4065e6722e08-images\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047795 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwznh\" (UniqueName: \"kubernetes.io/projected/cafb5284-e531-4979-86aa-ff4c626b71be-kube-api-access-nwznh\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047824 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56fe8e2e-1586-4fbd-b68b-f207813a7c35-service-ca-bundle\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047863 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a3aa4249-2022-45e7-844a-25e6b14f0661-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047920 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8afa818f-4048-49d9-a54a-3ba48f3de8cd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047953 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8afa818f-4048-49d9-a54a-3ba48f3de8cd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.047981 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e87e9694-6803-4745-8a15-4065e6722e08-auth-proxy-config\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.048017 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-metrics-certs\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.048056 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t59mq\" (UniqueName: \"kubernetes.io/projected/7ef124c8-9694-4320-bdda-2ecc10489225-kube-api-access-t59mq\") pod \"dns-operator-744455d44c-nt44m\" (UID: \"7ef124c8-9694-4320-bdda-2ecc10489225\") " pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.048095 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/670f4f60-09f3-4a59-bc86-b34fe9466e2e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.048856 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-config\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.049117 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8afa818f-4048-49d9-a54a-3ba48f3de8cd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.049384 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-ca\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.049515 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-service-ca\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.050773 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a3aa4249-2022-45e7-844a-25e6b14f0661-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.051230 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ef124c8-9694-4320-bdda-2ecc10489225-metrics-tls\") pod \"dns-operator-744455d44c-nt44m\" (UID: \"7ef124c8-9694-4320-bdda-2ecc10489225\") " pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.051232 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8afa818f-4048-49d9-a54a-3ba48f3de8cd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.052816 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cafb5284-e531-4979-86aa-ff4c626b71be-serving-cert\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.053516 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cafb5284-e531-4979-86aa-ff4c626b71be-etcd-client\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.066646 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.086238 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.105995 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.126479 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.145922 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.173230 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.186798 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.207075 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.226659 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.247172 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.266392 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.286808 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.307932 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.327711 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.348217 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.367490 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.387042 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.390198 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e87e9694-6803-4745-8a15-4065e6722e08-images\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.406810 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.427428 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.432437 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e87e9694-6803-4745-8a15-4065e6722e08-proxy-tls\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.447945 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.466484 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.473080 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1b905b7b-c2ff-4cf1-9102-d7124d618ec7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fsx6z\" (UID: \"1b905b7b-c2ff-4cf1-9102-d7124d618ec7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.488459 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.496976 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a3aa4249-2022-45e7-844a-25e6b14f0661-proxy-tls\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.508938 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.527818 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.547824 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.567331 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.587815 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.607104 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.612053 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/670f4f60-09f3-4a59-bc86-b34fe9466e2e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.627793 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.630511 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/670f4f60-09f3-4a59-bc86-b34fe9466e2e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.656898 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.668897 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.687343 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.707154 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.714735 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-default-certificate\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.726915 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.733029 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-stats-auth\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.747949 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.753529 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56fe8e2e-1586-4fbd-b68b-f207813a7c35-metrics-certs\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.764878 4946 request.go:700] Waited for 1.016749335s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-dockercfg-zdk86&limit=500&resourceVersion=0 Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.767084 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.786604 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.790270 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56fe8e2e-1586-4fbd-b68b-f207813a7c35-service-ca-bundle\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.807358 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.828261 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.899161 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjpb6\" (UniqueName: \"kubernetes.io/projected/6c199c97-422e-4a3d-a598-f28a651b377f-kube-api-access-zjpb6\") pod \"machine-approver-56656f9798-rqm9w\" (UID: \"6c199c97-422e-4a3d-a598-f28a651b377f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.915526 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c79v5\" (UniqueName: \"kubernetes.io/projected/6ff611ff-6a55-435a-9723-df1b38741952-kube-api-access-c79v5\") pod \"machine-api-operator-5694c8668f-vcrrg\" (UID: \"6ff611ff-6a55-435a-9723-df1b38741952\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.927243 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.947071 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.967469 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 06:52:33 crc kubenswrapper[4946]: I1203 06:52:33.988331 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.007857 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.027265 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.046933 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.066773 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.087720 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.093100 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.106636 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.110082 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.126957 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.149286 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.166473 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.186428 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.207325 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.227548 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.230450 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" event={"ID":"6c199c97-422e-4a3d-a598-f28a651b377f","Type":"ContainerStarted","Data":"177f7198b885ab9f3cf72acdb6983a6c55844082741641115a540a827d2c789c"} Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.253088 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.267162 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.286408 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.295858 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vcrrg"] Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.309035 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.326560 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.346147 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.366882 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.386769 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.406866 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.427239 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.446822 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.467313 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.469842 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:34 crc kubenswrapper[4946]: E1203 06:52:34.470119 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:54:36.470075148 +0000 UTC m=+269.266765257 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.486597 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.507065 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.526280 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.547961 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.567373 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.571692 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.571832 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.571980 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.572109 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.579609 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.579784 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.580230 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.587194 4946 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.607877 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.628515 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.633612 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.647310 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.666761 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.706571 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtlvs\" (UniqueName: \"kubernetes.io/projected/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-kube-api-access-qtlvs\") pod \"route-controller-manager-6576b87f9c-kmnwr\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.714061 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.731825 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6jpb\" (UniqueName: \"kubernetes.io/projected/6c970075-0d42-45b3-a56f-f0686e7a2455-kube-api-access-d6jpb\") pod \"authentication-operator-69f744f599-f588s\" (UID: \"6c970075-0d42-45b3-a56f-f0686e7a2455\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.749610 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qms6c\" (UniqueName: \"kubernetes.io/projected/84f01e09-8152-4a65-b71a-5456ea557a77-kube-api-access-qms6c\") pod \"openshift-config-operator-7777fb866f-f59l8\" (UID: \"84f01e09-8152-4a65-b71a-5456ea557a77\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.765026 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjtdz\" (UniqueName: \"kubernetes.io/projected/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-kube-api-access-gjtdz\") pod \"controller-manager-879f6c89f-82rd5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.765108 4946 request.go:700] Waited for 1.814518316s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.779663 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.788245 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blgxs\" (UniqueName: \"kubernetes.io/projected/f2d67105-8c6c-44b1-9baa-71424ee41402-kube-api-access-blgxs\") pod \"oauth-openshift-558db77b4-s7slv\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.796475 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.800932 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.804901 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25c62\" (UniqueName: \"kubernetes.io/projected/f27b8b97-2934-4d16-a756-8d402ec86bad-kube-api-access-25c62\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.825179 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6nk5\" (UniqueName: \"kubernetes.io/projected/ca8d6db9-e160-4ad7-a399-3dc1618add0f-kube-api-access-m6nk5\") pod \"downloads-7954f5f757-ncr4w\" (UID: \"ca8d6db9-e160-4ad7-a399-3dc1618add0f\") " pod="openshift-console/downloads-7954f5f757-ncr4w" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.847146 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f27b8b97-2934-4d16-a756-8d402ec86bad-bound-sa-token\") pod \"ingress-operator-5b745b69d9-phv7g\" (UID: \"f27b8b97-2934-4d16-a756-8d402ec86bad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.854528 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.885919 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hckq6\" (UniqueName: \"kubernetes.io/projected/b43afb98-1ebd-4df2-85ea-c3c12c66ded1-kube-api-access-hckq6\") pod \"console-operator-58897d9998-w6x6n\" (UID: \"b43afb98-1ebd-4df2-85ea-c3c12c66ded1\") " pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.889319 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gwdg\" (UniqueName: \"kubernetes.io/projected/cf1e2d1c-6794-48f0-9dde-8eacb6a16525-kube-api-access-6gwdg\") pod \"cluster-samples-operator-665b6dd947-rcb2j\" (UID: \"cf1e2d1c-6794-48f0-9dde-8eacb6a16525\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.915431 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gssj8\" (UniqueName: \"kubernetes.io/projected/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-kube-api-access-gssj8\") pod \"console-f9d7485db-zkld2\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.922361 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.928668 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjm4q\" (UniqueName: \"kubernetes.io/projected/4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599-kube-api-access-jjm4q\") pod \"apiserver-76f77b778f-7nc8s\" (UID: \"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599\") " pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.953225 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9499\" (UniqueName: \"kubernetes.io/projected/ece5ff47-76df-4fad-9a59-3dbba27b788d-kube-api-access-c9499\") pod \"openshift-controller-manager-operator-756b6f6bc6-gfz9s\" (UID: \"ece5ff47-76df-4fad-9a59-3dbba27b788d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.963113 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ncr4w" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.975082 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkcxz\" (UniqueName: \"kubernetes.io/projected/f1bd7b04-110e-45fb-b972-c662a2e7c791-kube-api-access-tkcxz\") pod \"apiserver-7bbb656c7d-7hmqc\" (UID: \"f1bd7b04-110e-45fb-b972-c662a2e7c791\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.977644 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.993611 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqckd\" (UniqueName: \"kubernetes.io/projected/4a4c7b48-79a1-4a55-9012-a50c96908e7a-kube-api-access-xqckd\") pod \"openshift-apiserver-operator-796bbdcf4f-ghqsn\" (UID: \"4a4c7b48-79a1-4a55-9012-a50c96908e7a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:34 crc kubenswrapper[4946]: I1203 06:52:34.994004 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.010113 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkfpf\" (UniqueName: \"kubernetes.io/projected/a3aa4249-2022-45e7-844a-25e6b14f0661-kube-api-access-mkfpf\") pod \"machine-config-controller-84d6567774-2bq9b\" (UID: \"a3aa4249-2022-45e7-844a-25e6b14f0661\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.025203 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgtlw\" (UniqueName: \"kubernetes.io/projected/56fe8e2e-1586-4fbd-b68b-f207813a7c35-kube-api-access-rgtlw\") pod \"router-default-5444994796-l68r8\" (UID: \"56fe8e2e-1586-4fbd-b68b-f207813a7c35\") " pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.029322 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.035922 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.041758 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.069412 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f588s"] Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.078713 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l58b7\" (UniqueName: \"kubernetes.io/projected/8afa818f-4048-49d9-a54a-3ba48f3de8cd-kube-api-access-l58b7\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.081235 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/670f4f60-09f3-4a59-bc86-b34fe9466e2e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-698gl\" (UID: \"670f4f60-09f3-4a59-bc86-b34fe9466e2e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.082104 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t59mq\" (UniqueName: \"kubernetes.io/projected/7ef124c8-9694-4320-bdda-2ecc10489225-kube-api-access-t59mq\") pod \"dns-operator-744455d44c-nt44m\" (UID: \"7ef124c8-9694-4320-bdda-2ecc10489225\") " pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.088548 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.094291 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.094428 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.102963 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwznh\" (UniqueName: \"kubernetes.io/projected/cafb5284-e531-4979-86aa-ff4c626b71be-kube-api-access-nwznh\") pod \"etcd-operator-b45778765-qwfgf\" (UID: \"cafb5284-e531-4979-86aa-ff4c626b71be\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.110193 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.116589 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.123802 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plgvs\" (UniqueName: \"kubernetes.io/projected/1b905b7b-c2ff-4cf1-9102-d7124d618ec7-kube-api-access-plgvs\") pod \"multus-admission-controller-857f4d67dd-fsx6z\" (UID: \"1b905b7b-c2ff-4cf1-9102-d7124d618ec7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.135420 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.143432 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdh4p\" (UniqueName: \"kubernetes.io/projected/e87e9694-6803-4745-8a15-4065e6722e08-kube-api-access-fdh4p\") pod \"machine-config-operator-74547568cd-drm7q\" (UID: \"e87e9694-6803-4745-8a15-4065e6722e08\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.152681 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr"] Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.153173 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.159326 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.163344 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8afa818f-4048-49d9-a54a-3ba48f3de8cd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b926l\" (UID: \"8afa818f-4048-49d9-a54a-3ba48f3de8cd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.166047 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.174469 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.185183 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-s7slv"] Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.189528 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.232626 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.245011 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6be31e1e9990c80e7e889e53d670003f90b8b7da944b08fbac546766de52f4fd"} Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.246296 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"317d628811bd730aa0a223c2d7185bab3f4c0ff93c92c4d070d8623a61451054"} Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.246348 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"bca2e86dfcd1c7eeb1655c946deb34e763877060a5b05177d75cb6bb1e33f1da"} Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.246986 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.250999 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" event={"ID":"6ff611ff-6a55-435a-9723-df1b38741952","Type":"ContainerStarted","Data":"abe5a5d0dd4fd73c0ca2137a0eacd1b428b34703f6a670f5450f795ccebad2c6"} Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.251029 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" event={"ID":"6ff611ff-6a55-435a-9723-df1b38741952","Type":"ContainerStarted","Data":"542f8d2773495f776646e2464222f1f636b504c7b98193b85cfaef6388319660"} Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.251040 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" event={"ID":"6ff611ff-6a55-435a-9723-df1b38741952","Type":"ContainerStarted","Data":"51e01a653b710234351eb828c9b132af0ce67199ca30f22836c71b6d4d3a9c07"} Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.255753 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" event={"ID":"6c199c97-422e-4a3d-a598-f28a651b377f","Type":"ContainerStarted","Data":"8732b8e6034032cd8e38b68125272140fc6df07f048d6086c58d4a522a186b31"} Dec 03 06:52:35 crc kubenswrapper[4946]: W1203 06:52:35.260551 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf86bb6c2_c6d0_4c6a_9801_d1d8f34a4e16.slice/crio-b4fc0751b413e63b6a4547399c2d87d0e5396e84af3ff4264358efc6d836d04d WatchSource:0}: Error finding container b4fc0751b413e63b6a4547399c2d87d0e5396e84af3ff4264358efc6d836d04d: Status 404 returned error can't find the container with id b4fc0751b413e63b6a4547399c2d87d0e5396e84af3ff4264358efc6d836d04d Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.261005 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" event={"ID":"6c970075-0d42-45b3-a56f-f0686e7a2455","Type":"ContainerStarted","Data":"6d04f22c8e966d8d360d7cff98e824714a6d2012e8f2e6f0718b43419bbebd3e"} Dec 03 06:52:35 crc kubenswrapper[4946]: W1203 06:52:35.270412 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2d67105_8c6c_44b1_9baa_71424ee41402.slice/crio-e5796b5e54679eae5273ab9df7a29d204f5b45c5a2ec8b20b7591a353c5594f5 WatchSource:0}: Error finding container e5796b5e54679eae5273ab9df7a29d204f5b45c5a2ec8b20b7591a353c5594f5: Status 404 returned error can't find the container with id e5796b5e54679eae5273ab9df7a29d204f5b45c5a2ec8b20b7591a353c5594f5 Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.283539 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6050037f-7b92-4a14-8597-3b6be013e8c9-config\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.283593 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.283621 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.283649 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6ls6\" (UniqueName: \"kubernetes.io/projected/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-kube-api-access-s6ls6\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.283716 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8532cd3f-9a90-42cd-b237-44e18b9117aa-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284252 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6050037f-7b92-4a14-8597-3b6be013e8c9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: E1203 06:52:35.284276 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:35.784253289 +0000 UTC m=+148.580943398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284326 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6050037f-7b92-4a14-8597-3b6be013e8c9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284429 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-certificates\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284524 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47hmk\" (UniqueName: \"kubernetes.io/projected/8532cd3f-9a90-42cd-b237-44e18b9117aa-kube-api-access-47hmk\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284579 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284731 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgxqp\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-kube-api-access-xgxqp\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284816 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284858 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-config\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284898 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b8c7afc-56fb-49ab-ae82-e60167809f15-ca-trust-extracted\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284940 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-trusted-ca\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.284981 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b8c7afc-56fb-49ab-ae82-e60167809f15-installation-pull-secrets\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.285057 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-tls\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.285170 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-bound-sa-token\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.285515 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8532cd3f-9a90-42cd-b237-44e18b9117aa-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.285592 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: W1203 06:52:35.383580 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-9063a74a700ad2399bc67139789184d3e094d1e98ce526c28421196fdcb05f8f WatchSource:0}: Error finding container 9063a74a700ad2399bc67139789184d3e094d1e98ce526c28421196fdcb05f8f: Status 404 returned error can't find the container with id 9063a74a700ad2399bc67139789184d3e094d1e98ce526c28421196fdcb05f8f Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.390860 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:35 crc kubenswrapper[4946]: E1203 06:52:35.391239 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:35.891212958 +0000 UTC m=+148.687903067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.391343 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv6xr\" (UniqueName: \"kubernetes.io/projected/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-kube-api-access-pv6xr\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.391371 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c58dec3d-f028-43b2-8e71-1954da44b01c-config-volume\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.391623 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8532cd3f-9a90-42cd-b237-44e18b9117aa-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.391680 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lghxw\" (UniqueName: \"kubernetes.io/projected/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-kube-api-access-lghxw\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.392535 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8532cd3f-9a90-42cd-b237-44e18b9117aa-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.392617 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.392665 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4bl4\" (UniqueName: \"kubernetes.io/projected/cbc2e507-308f-4be7-af6c-c356d3e7d2c5-kube-api-access-p4bl4\") pod \"ingress-canary-qsg2m\" (UID: \"cbc2e507-308f-4be7-af6c-c356d3e7d2c5\") " pod="openshift-ingress-canary/ingress-canary-qsg2m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.392726 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6050037f-7b92-4a14-8597-3b6be013e8c9-config\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.392768 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cbc2e507-308f-4be7-af6c-c356d3e7d2c5-cert\") pod \"ingress-canary-qsg2m\" (UID: \"cbc2e507-308f-4be7-af6c-c356d3e7d2c5\") " pod="openshift-ingress-canary/ingress-canary-qsg2m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.392810 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-socket-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.392841 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.392866 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: E1203 06:52:35.393116 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:35.893105932 +0000 UTC m=+148.689796041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.394011 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-node-bootstrap-token\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.394109 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6ls6\" (UniqueName: \"kubernetes.io/projected/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-kube-api-access-s6ls6\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.394138 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-apiservice-cert\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.394968 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/33e1bf8e-409b-4546-9eaa-c6099871e3d2-signing-cabundle\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.395122 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8532cd3f-9a90-42cd-b237-44e18b9117aa-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.395391 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6050037f-7b92-4a14-8597-3b6be013e8c9-config\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.396637 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6050037f-7b92-4a14-8597-3b6be013e8c9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.397404 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86gct\" (UniqueName: \"kubernetes.io/projected/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-kube-api-access-86gct\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.397468 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6050037f-7b92-4a14-8597-3b6be013e8c9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.397488 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pj2t\" (UniqueName: \"kubernetes.io/projected/33e1bf8e-409b-4546-9eaa-c6099871e3d2-kube-api-access-9pj2t\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.397517 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-plugins-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.397553 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n24z\" (UniqueName: \"kubernetes.io/projected/0acb31c7-623c-40b7-889c-4a9a45b55cf1-kube-api-access-7n24z\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5trw\" (UID: \"0acb31c7-623c-40b7-889c-4a9a45b55cf1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.397595 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-certs\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.397614 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmmvs\" (UniqueName: \"kubernetes.io/projected/1547b1af-d1e0-44a9-9831-30c901233123-kube-api-access-tmmvs\") pod \"package-server-manager-789f6589d5-96d5q\" (UID: \"1547b1af-d1e0-44a9-9831-30c901233123\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.397630 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-srv-cert\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400068 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-certificates\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400126 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47hmk\" (UniqueName: \"kubernetes.io/projected/8532cd3f-9a90-42cd-b237-44e18b9117aa-kube-api-access-47hmk\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400251 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400280 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxk77\" (UniqueName: \"kubernetes.io/projected/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-kube-api-access-dxk77\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400307 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-metrics-tls\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400329 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-tmpfs\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400378 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/554d7c42-1f14-4bac-b062-bdb816f18789-config\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400419 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4pxz\" (UniqueName: \"kubernetes.io/projected/09b65f3a-96c6-4342-802d-8f83502aee5d-kube-api-access-q4pxz\") pod \"migrator-59844c95c7-s9927\" (UID: \"09b65f3a-96c6-4342-802d-8f83502aee5d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400450 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2j8j\" (UniqueName: \"kubernetes.io/projected/4254dd3d-9a3f-475a-855b-95f7101c7ab9-kube-api-access-r2j8j\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400488 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c58dec3d-f028-43b2-8e71-1954da44b01c-secret-volume\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400512 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0acb31c7-623c-40b7-889c-4a9a45b55cf1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5trw\" (UID: \"0acb31c7-623c-40b7-889c-4a9a45b55cf1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400587 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgxqp\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-kube-api-access-xgxqp\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400606 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-config-volume\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400620 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-webhook-cert\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400640 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-mountpoint-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400720 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400770 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4254dd3d-9a3f-475a-855b-95f7101c7ab9-srv-cert\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400838 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-config\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400861 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-598dd\" (UniqueName: \"kubernetes.io/projected/554d7c42-1f14-4bac-b062-bdb816f18789-kube-api-access-598dd\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400928 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b8c7afc-56fb-49ab-ae82-e60167809f15-ca-trust-extracted\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.400961 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxw9d\" (UniqueName: \"kubernetes.io/projected/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-kube-api-access-lxw9d\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401015 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-csi-data-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401052 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-trusted-ca\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401080 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401130 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-registration-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401155 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/554d7c42-1f14-4bac-b062-bdb816f18789-serving-cert\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401182 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/33e1bf8e-409b-4546-9eaa-c6099871e3d2-signing-key\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401203 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngm49\" (UniqueName: \"kubernetes.io/projected/c58dec3d-f028-43b2-8e71-1954da44b01c-kube-api-access-ngm49\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401237 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b8c7afc-56fb-49ab-ae82-e60167809f15-installation-pull-secrets\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401278 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-tls\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401302 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1547b1af-d1e0-44a9-9831-30c901233123-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-96d5q\" (UID: \"1547b1af-d1e0-44a9-9831-30c901233123\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401360 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4254dd3d-9a3f-475a-855b-95f7101c7ab9-profile-collector-cert\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.401383 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-bound-sa-token\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.403000 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6050037f-7b92-4a14-8597-3b6be013e8c9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.403309 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.404096 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-certificates\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.404285 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-config\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.404999 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.406343 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b8c7afc-56fb-49ab-ae82-e60167809f15-ca-trust-extracted\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.406615 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8532cd3f-9a90-42cd-b237-44e18b9117aa-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.407461 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.407517 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-trusted-ca\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.409397 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.414371 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b8c7afc-56fb-49ab-ae82-e60167809f15-installation-pull-secrets\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.437372 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-tls\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.448571 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6ls6\" (UniqueName: \"kubernetes.io/projected/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-kube-api-access-s6ls6\") pod \"marketplace-operator-79b997595-gzrcc\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.468291 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6050037f-7b92-4a14-8597-3b6be013e8c9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f6b24\" (UID: \"6050037f-7b92-4a14-8597-3b6be013e8c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.486139 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.490029 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47hmk\" (UniqueName: \"kubernetes.io/projected/8532cd3f-9a90-42cd-b237-44e18b9117aa-kube-api-access-47hmk\") pod \"kube-storage-version-migrator-operator-b67b599dd-gzfs9\" (UID: \"8532cd3f-9a90-42cd-b237-44e18b9117aa\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502072 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502231 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-config-volume\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502262 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-webhook-cert\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502286 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-mountpoint-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502316 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4254dd3d-9a3f-475a-855b-95f7101c7ab9-srv-cert\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502335 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-598dd\" (UniqueName: \"kubernetes.io/projected/554d7c42-1f14-4bac-b062-bdb816f18789-kube-api-access-598dd\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502362 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxw9d\" (UniqueName: \"kubernetes.io/projected/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-kube-api-access-lxw9d\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502384 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502410 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-csi-data-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502429 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-registration-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502448 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/554d7c42-1f14-4bac-b062-bdb816f18789-serving-cert\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502468 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/33e1bf8e-409b-4546-9eaa-c6099871e3d2-signing-key\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502492 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngm49\" (UniqueName: \"kubernetes.io/projected/c58dec3d-f028-43b2-8e71-1954da44b01c-kube-api-access-ngm49\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502517 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1547b1af-d1e0-44a9-9831-30c901233123-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-96d5q\" (UID: \"1547b1af-d1e0-44a9-9831-30c901233123\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502540 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4254dd3d-9a3f-475a-855b-95f7101c7ab9-profile-collector-cert\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502564 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c58dec3d-f028-43b2-8e71-1954da44b01c-config-volume\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502580 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv6xr\" (UniqueName: \"kubernetes.io/projected/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-kube-api-access-pv6xr\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502598 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lghxw\" (UniqueName: \"kubernetes.io/projected/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-kube-api-access-lghxw\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502616 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4bl4\" (UniqueName: \"kubernetes.io/projected/cbc2e507-308f-4be7-af6c-c356d3e7d2c5-kube-api-access-p4bl4\") pod \"ingress-canary-qsg2m\" (UID: \"cbc2e507-308f-4be7-af6c-c356d3e7d2c5\") " pod="openshift-ingress-canary/ingress-canary-qsg2m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502633 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cbc2e507-308f-4be7-af6c-c356d3e7d2c5-cert\") pod \"ingress-canary-qsg2m\" (UID: \"cbc2e507-308f-4be7-af6c-c356d3e7d2c5\") " pod="openshift-ingress-canary/ingress-canary-qsg2m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502651 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-socket-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502676 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-node-bootstrap-token\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502692 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-apiservice-cert\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502713 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/33e1bf8e-409b-4546-9eaa-c6099871e3d2-signing-cabundle\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502779 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pj2t\" (UniqueName: \"kubernetes.io/projected/33e1bf8e-409b-4546-9eaa-c6099871e3d2-kube-api-access-9pj2t\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502800 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86gct\" (UniqueName: \"kubernetes.io/projected/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-kube-api-access-86gct\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502815 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-plugins-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502831 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n24z\" (UniqueName: \"kubernetes.io/projected/0acb31c7-623c-40b7-889c-4a9a45b55cf1-kube-api-access-7n24z\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5trw\" (UID: \"0acb31c7-623c-40b7-889c-4a9a45b55cf1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502845 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-certs\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502859 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmmvs\" (UniqueName: \"kubernetes.io/projected/1547b1af-d1e0-44a9-9831-30c901233123-kube-api-access-tmmvs\") pod \"package-server-manager-789f6589d5-96d5q\" (UID: \"1547b1af-d1e0-44a9-9831-30c901233123\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502873 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-srv-cert\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502905 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxk77\" (UniqueName: \"kubernetes.io/projected/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-kube-api-access-dxk77\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502919 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-tmpfs\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502935 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/554d7c42-1f14-4bac-b062-bdb816f18789-config\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502950 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4pxz\" (UniqueName: \"kubernetes.io/projected/09b65f3a-96c6-4342-802d-8f83502aee5d-kube-api-access-q4pxz\") pod \"migrator-59844c95c7-s9927\" (UID: \"09b65f3a-96c6-4342-802d-8f83502aee5d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502966 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-metrics-tls\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502981 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2j8j\" (UniqueName: \"kubernetes.io/projected/4254dd3d-9a3f-475a-855b-95f7101c7ab9-kube-api-access-r2j8j\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.502996 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c58dec3d-f028-43b2-8e71-1954da44b01c-secret-volume\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.503011 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0acb31c7-623c-40b7-889c-4a9a45b55cf1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5trw\" (UID: \"0acb31c7-623c-40b7-889c-4a9a45b55cf1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" Dec 03 06:52:35 crc kubenswrapper[4946]: E1203 06:52:35.504100 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.004072786 +0000 UTC m=+148.800762955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.505065 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-tmpfs\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.505691 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/554d7c42-1f14-4bac-b062-bdb816f18789-config\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.505700 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-mountpoint-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.505774 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-socket-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.506020 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-csi-data-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.506168 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgxqp\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-kube-api-access-xgxqp\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.506284 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-registration-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.507294 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-plugins-dir\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.507507 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-config-volume\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.507933 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/33e1bf8e-409b-4546-9eaa-c6099871e3d2-signing-cabundle\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.508920 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cbc2e507-308f-4be7-af6c-c356d3e7d2c5-cert\") pod \"ingress-canary-qsg2m\" (UID: \"cbc2e507-308f-4be7-af6c-c356d3e7d2c5\") " pod="openshift-ingress-canary/ingress-canary-qsg2m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.510458 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c58dec3d-f028-43b2-8e71-1954da44b01c-config-volume\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.510528 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-metrics-tls\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.510623 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/554d7c42-1f14-4bac-b062-bdb816f18789-serving-cert\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.511716 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zkld2"] Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.512659 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/33e1bf8e-409b-4546-9eaa-c6099871e3d2-signing-key\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.514042 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.514102 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4254dd3d-9a3f-475a-855b-95f7101c7ab9-profile-collector-cert\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.514863 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4254dd3d-9a3f-475a-855b-95f7101c7ab9-srv-cert\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.521836 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-webhook-cert\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.523756 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-apiservice-cert\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.523762 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0acb31c7-623c-40b7-889c-4a9a45b55cf1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5trw\" (UID: \"0acb31c7-623c-40b7-889c-4a9a45b55cf1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.523880 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c58dec3d-f028-43b2-8e71-1954da44b01c-secret-volume\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.524371 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-srv-cert\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.524673 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-node-bootstrap-token\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.525597 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-certs\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.540956 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1547b1af-d1e0-44a9-9831-30c901233123-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-96d5q\" (UID: \"1547b1af-d1e0-44a9-9831-30c901233123\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.541576 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ec05e39-71b1-46b8-a6e4-8312e4d6d020-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9spfp\" (UID: \"7ec05e39-71b1-46b8-a6e4-8312e4d6d020\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.554089 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-bound-sa-token\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.589149 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j"] Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.598875 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4bl4\" (UniqueName: \"kubernetes.io/projected/cbc2e507-308f-4be7-af6c-c356d3e7d2c5-kube-api-access-p4bl4\") pod \"ingress-canary-qsg2m\" (UID: \"cbc2e507-308f-4be7-af6c-c356d3e7d2c5\") " pod="openshift-ingress-canary/ingress-canary-qsg2m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.605988 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.606333 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmmvs\" (UniqueName: \"kubernetes.io/projected/1547b1af-d1e0-44a9-9831-30c901233123-kube-api-access-tmmvs\") pod \"package-server-manager-789f6589d5-96d5q\" (UID: \"1547b1af-d1e0-44a9-9831-30c901233123\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:35 crc kubenswrapper[4946]: E1203 06:52:35.606884 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.106859556 +0000 UTC m=+148.903549665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.628100 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4pxz\" (UniqueName: \"kubernetes.io/projected/09b65f3a-96c6-4342-802d-8f83502aee5d-kube-api-access-q4pxz\") pod \"migrator-59844c95c7-s9927\" (UID: \"09b65f3a-96c6-4342-802d-8f83502aee5d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.649112 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxk77\" (UniqueName: \"kubernetes.io/projected/cf3da82c-a93f-42ae-8935-b6dba92d2cb5-kube-api-access-dxk77\") pod \"machine-config-server-8rwkt\" (UID: \"cf3da82c-a93f-42ae-8935-b6dba92d2cb5\") " pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.653037 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ncr4w"] Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.678805 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2j8j\" (UniqueName: \"kubernetes.io/projected/4254dd3d-9a3f-475a-855b-95f7101c7ab9-kube-api-access-r2j8j\") pod \"catalog-operator-68c6474976-fgbb8\" (UID: \"4254dd3d-9a3f-475a-855b-95f7101c7ab9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.689583 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxw9d\" (UniqueName: \"kubernetes.io/projected/bacf1b15-c2aa-4d79-8bc9-2977deea4ef7-kube-api-access-lxw9d\") pod \"dns-default-2q2qg\" (UID: \"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7\") " pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.703470 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngm49\" (UniqueName: \"kubernetes.io/projected/c58dec3d-f028-43b2-8e71-1954da44b01c-kube-api-access-ngm49\") pod \"collect-profiles-29412405-8gzbd\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.707275 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:35 crc kubenswrapper[4946]: E1203 06:52:35.707988 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.207715771 +0000 UTC m=+149.004405880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.727054 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.742041 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.747731 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.752366 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pj2t\" (UniqueName: \"kubernetes.io/projected/33e1bf8e-409b-4546-9eaa-c6099871e3d2-kube-api-access-9pj2t\") pod \"service-ca-9c57cc56f-7sng8\" (UID: \"33e1bf8e-409b-4546-9eaa-c6099871e3d2\") " pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.753313 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-598dd\" (UniqueName: \"kubernetes.io/projected/554d7c42-1f14-4bac-b062-bdb816f18789-kube-api-access-598dd\") pod \"service-ca-operator-777779d784-wr8vk\" (UID: \"554d7c42-1f14-4bac-b062-bdb816f18789\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.796078 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.797591 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86gct\" (UniqueName: \"kubernetes.io/projected/f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9-kube-api-access-86gct\") pod \"csi-hostpathplugin-55ddd\" (UID: \"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9\") " pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.805557 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv6xr\" (UniqueName: \"kubernetes.io/projected/f4bf5842-4786-458d-9cdd-8f77b5bbce1f-kube-api-access-pv6xr\") pod \"olm-operator-6b444d44fb-xtwj9\" (UID: \"f4bf5842-4786-458d-9cdd-8f77b5bbce1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.807602 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.809088 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:35 crc kubenswrapper[4946]: E1203 06:52:35.809486 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.309470342 +0000 UTC m=+149.106160451 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.810303 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.816986 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.820800 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n24z\" (UniqueName: \"kubernetes.io/projected/0acb31c7-623c-40b7-889c-4a9a45b55cf1-kube-api-access-7n24z\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5trw\" (UID: \"0acb31c7-623c-40b7-889c-4a9a45b55cf1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.824161 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.838180 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.842173 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lghxw\" (UniqueName: \"kubernetes.io/projected/9a05ab86-ca6d-418d-a3a7-cd8f958bca06-kube-api-access-lghxw\") pod \"packageserver-d55dfcdfc-c4xq7\" (UID: \"9a05ab86-ca6d-418d-a3a7-cd8f958bca06\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.845125 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.851725 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8rwkt" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.863351 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.880968 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qsg2m" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.892280 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-55ddd" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.900901 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:35 crc kubenswrapper[4946]: I1203 06:52:35.910064 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:35 crc kubenswrapper[4946]: E1203 06:52:35.910648 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.410627826 +0000 UTC m=+149.207317935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.011503 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.012195 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.51217913 +0000 UTC m=+149.308869239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.117016 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.117380 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.617357639 +0000 UTC m=+149.414047748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.130710 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.218166 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.218569 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.718527243 +0000 UTC m=+149.515217342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.293423 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" event={"ID":"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16","Type":"ContainerStarted","Data":"2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.293479 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" event={"ID":"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16","Type":"ContainerStarted","Data":"b4fc0751b413e63b6a4547399c2d87d0e5396e84af3ff4264358efc6d836d04d"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.294026 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.295467 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" event={"ID":"6c970075-0d42-45b3-a56f-f0686e7a2455","Type":"ContainerStarted","Data":"9a7c6cabc97cbf5c0abcfd630a5909a52297902aa2886482fb0cc98be24f8622"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.296750 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f59l8"] Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.312442 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fe2bc3d56f98747d5a69d7d9804b778fb91e9719e03076bd51fc79233496da29"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.314708 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" event={"ID":"f2d67105-8c6c-44b1-9baa-71424ee41402","Type":"ContainerStarted","Data":"e5796b5e54679eae5273ab9df7a29d204f5b45c5a2ec8b20b7591a353c5594f5"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.317984 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-l68r8" event={"ID":"56fe8e2e-1586-4fbd-b68b-f207813a7c35","Type":"ContainerStarted","Data":"65e5be22daca08b06de717f002fd3023dd99a2fd764c20fd37ef7ac8f160b1a0"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.318014 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-l68r8" event={"ID":"56fe8e2e-1586-4fbd-b68b-f207813a7c35","Type":"ContainerStarted","Data":"8c901a816a2098c9eecc6f303c09852a3c97322e249aa9c2b9efee6fd3e382dc"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.319147 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.319640 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.819620645 +0000 UTC m=+149.616310754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.342516 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-82rd5"] Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.356998 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkld2" event={"ID":"a89bae42-673b-4eb1-87d2-ef2dd919c4d4","Type":"ContainerStarted","Data":"b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.357075 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkld2" event={"ID":"a89bae42-673b-4eb1-87d2-ef2dd919c4d4","Type":"ContainerStarted","Data":"bf44b5944aec0bb6730f8813277691ed0968ac0e2d0da5c4ed06ef9b5f012c53"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.398184 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ncr4w" event={"ID":"ca8d6db9-e160-4ad7-a399-3dc1618add0f","Type":"ContainerStarted","Data":"8b50a8719083507f2c0928c421c5afcadee9027aa1f6601f127fe81878e1bf2b"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.398838 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ncr4w" Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.404049 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2661246db5d067653d9e21b05f467155c7421d8ac36ddff082571a1cc9eea1b1"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.404090 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9063a74a700ad2399bc67139789184d3e094d1e98ce526c28421196fdcb05f8f"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.407080 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8rwkt" event={"ID":"cf3da82c-a93f-42ae-8935-b6dba92d2cb5","Type":"ContainerStarted","Data":"fbfe4678615a68caf9c7790414432628e12c84b8f1cf26bbee873514050aa77b"} Dec 03 06:52:36 crc kubenswrapper[4946]: W1203 06:52:36.411016 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84f01e09_8152_4a65_b71a_5456ea557a77.slice/crio-53bd08006d98e1ca92757af8c690e11e1ca138bdc2481c3f97147d453aa13901 WatchSource:0}: Error finding container 53bd08006d98e1ca92757af8c690e11e1ca138bdc2481c3f97147d453aa13901: Status 404 returned error can't find the container with id 53bd08006d98e1ca92757af8c690e11e1ca138bdc2481c3f97147d453aa13901 Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.411805 4946 patch_prober.go:28] interesting pod/downloads-7954f5f757-ncr4w container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.411838 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ncr4w" podUID="ca8d6db9-e160-4ad7-a399-3dc1618add0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.418117 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" event={"ID":"6c199c97-422e-4a3d-a598-f28a651b377f","Type":"ContainerStarted","Data":"511093762c319d2744f7c27cc901607d2dbb7148f4cef887cfc6804199aa4d63"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.421190 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.422795 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:36.922783066 +0000 UTC m=+149.719473175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.422945 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" event={"ID":"cf1e2d1c-6794-48f0-9dde-8eacb6a16525","Type":"ContainerStarted","Data":"e80890b8440d836ad941d136cfa3d5797ff9352adfbcaac522f8d9d6327b40c4"} Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.453645 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-vcrrg" podStartSLOduration=123.453626449 podStartE2EDuration="2m3.453626449s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:36.453229007 +0000 UTC m=+149.249919116" watchObservedRunningTime="2025-12-03 06:52:36.453626449 +0000 UTC m=+149.250316558" Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.522509 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.524382 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.024365712 +0000 UTC m=+149.821055821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.624808 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.625098 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.125076573 +0000 UTC m=+149.921766802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.726049 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.726436 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.226402982 +0000 UTC m=+150.023093081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.827588 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.827956 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.327944117 +0000 UTC m=+150.124634226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.928125 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.928310 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.428270716 +0000 UTC m=+150.224960825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:36 crc kubenswrapper[4946]: I1203 06:52:36.928369 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:36 crc kubenswrapper[4946]: E1203 06:52:36.928883 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.428874144 +0000 UTC m=+150.225564243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.034275 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.034904 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.534727332 +0000 UTC m=+150.331417441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.106485 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.136300 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.136687 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.636671388 +0000 UTC m=+150.433361497 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.191256 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.238325 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.238959 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.738933883 +0000 UTC m=+150.535623992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.273756 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:37 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:37 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:37 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.274286 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.347602 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.348301 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.848282251 +0000 UTC m=+150.644972360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.357526 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w6x6n"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.382762 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.383921 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-l68r8" podStartSLOduration=125.38389247 podStartE2EDuration="2m5.38389247s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.331963564 +0000 UTC m=+150.128653673" watchObservedRunningTime="2025-12-03 06:52:37.38389247 +0000 UTC m=+150.180582579" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.413845 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nt44m"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.413937 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" podStartSLOduration=124.413893478 podStartE2EDuration="2m4.413893478s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.371809984 +0000 UTC m=+150.168500093" watchObservedRunningTime="2025-12-03 06:52:37.413893478 +0000 UTC m=+150.210583587" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.444027 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.449578 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.450053 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:37.950039192 +0000 UTC m=+150.746729301 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.450331 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" event={"ID":"f2d67105-8c6c-44b1-9baa-71424ee41402","Type":"ContainerStarted","Data":"dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.455669 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ncr4w" event={"ID":"ca8d6db9-e160-4ad7-a399-3dc1618add0f","Type":"ContainerStarted","Data":"bcda92e936eb9d616992653305c869634f45b6fa9f905844e47aee1165dea161"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.456639 4946 patch_prober.go:28] interesting pod/downloads-7954f5f757-ncr4w container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.456667 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ncr4w" podUID="ca8d6db9-e160-4ad7-a399-3dc1618add0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.469985 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" event={"ID":"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5","Type":"ContainerStarted","Data":"6af688b614e38e8e553cc6aed1dd1c83f1e9f9912bbe4fba19e1391109faec24"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.470020 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" event={"ID":"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5","Type":"ContainerStarted","Data":"68cd3d569acc945658276c6b4a5521c40c95690c60c7e49c388af7bcf6bfd58e"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.470923 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.488473 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" event={"ID":"7ef124c8-9694-4320-bdda-2ecc10489225","Type":"ContainerStarted","Data":"6022efd5cc311c5df0537489bac86a3dd6925ae5ff50f5c5fbc82fbe0c04b688"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.490173 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.506055 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ncr4w" podStartSLOduration=125.506014723 podStartE2EDuration="2m5.506014723s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.489669426 +0000 UTC m=+150.286359535" watchObservedRunningTime="2025-12-03 06:52:37.506014723 +0000 UTC m=+150.302704822" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.507256 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" event={"ID":"cf1e2d1c-6794-48f0-9dde-8eacb6a16525","Type":"ContainerStarted","Data":"e722c2b48db23b133a55acdd02f8ec900f899f7c8bc2e75ab75a5d1e5ddf8ed5"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.507469 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" event={"ID":"cf1e2d1c-6794-48f0-9dde-8eacb6a16525","Type":"ContainerStarted","Data":"23f2c5a6e0cafdc2ef692a605575286fe2b27ce018d51cb4bd4a43b3d06c67c1"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.525493 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" event={"ID":"ece5ff47-76df-4fad-9a59-3dbba27b788d","Type":"ContainerStarted","Data":"41b76cae6c99b7bcd2ded77fc17fe8bc547a6bf0934e7653472960834941a323"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.526411 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rqm9w" podStartSLOduration=125.526371256 podStartE2EDuration="2m5.526371256s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.52581035 +0000 UTC m=+150.322500479" watchObservedRunningTime="2025-12-03 06:52:37.526371256 +0000 UTC m=+150.323061375" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.542428 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w6x6n" event={"ID":"b43afb98-1ebd-4df2-85ea-c3c12c66ded1","Type":"ContainerStarted","Data":"13d0e28baec2914d449905d7a4174a0731be996350d8c4e9dbd72b652d6fe138"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.552493 4946 generic.go:334] "Generic (PLEG): container finished" podID="84f01e09-8152-4a65-b71a-5456ea557a77" containerID="253ed17dc17a8355597fb2fd4a9d1378f3acc040fe4c6bf3870ea7dcef3e2fef" exitCode=0 Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.552645 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" event={"ID":"84f01e09-8152-4a65-b71a-5456ea557a77","Type":"ContainerDied","Data":"253ed17dc17a8355597fb2fd4a9d1378f3acc040fe4c6bf3870ea7dcef3e2fef"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.552682 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" event={"ID":"84f01e09-8152-4a65-b71a-5456ea557a77","Type":"ContainerStarted","Data":"53bd08006d98e1ca92757af8c690e11e1ca138bdc2481c3f97147d453aa13901"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.553692 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.555382 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.055367685 +0000 UTC m=+150.852057794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.563204 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-f588s" podStartSLOduration=125.563188469 podStartE2EDuration="2m5.563188469s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.562122688 +0000 UTC m=+150.358812797" watchObservedRunningTime="2025-12-03 06:52:37.563188469 +0000 UTC m=+150.359878578" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.565914 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8rwkt" event={"ID":"cf3da82c-a93f-42ae-8935-b6dba92d2cb5","Type":"ContainerStarted","Data":"f9d3618498b09d6417279f598aa4e757fe3772dfae64f1a46496d6dbfba764df"} Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.609359 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-zkld2" podStartSLOduration=125.609333129 podStartE2EDuration="2m5.609333129s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.604998615 +0000 UTC m=+150.401688724" watchObservedRunningTime="2025-12-03 06:52:37.609333129 +0000 UTC m=+150.406023238" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.623213 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qwfgf"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.657425 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.658945 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.158922837 +0000 UTC m=+150.955613026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.662160 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" podStartSLOduration=125.662139529 podStartE2EDuration="2m5.662139529s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.652121433 +0000 UTC m=+150.448811542" watchObservedRunningTime="2025-12-03 06:52:37.662139529 +0000 UTC m=+150.458829628" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.663802 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.671540 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.672585 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.674476 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.174460812 +0000 UTC m=+150.971150911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: W1203 06:52:37.685363 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod670f4f60_09f3_4a59_bc86_b34fe9466e2e.slice/crio-aebfb118597562589392d091fe4377a0e86731a803ce0811fdf49d3dc67d66b5 WatchSource:0}: Error finding container aebfb118597562589392d091fe4377a0e86731a803ce0811fdf49d3dc67d66b5: Status 404 returned error can't find the container with id aebfb118597562589392d091fe4377a0e86731a803ce0811fdf49d3dc67d66b5 Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.695812 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8rwkt" podStartSLOduration=5.695787162 podStartE2EDuration="5.695787162s" podCreationTimestamp="2025-12-03 06:52:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.687136625 +0000 UTC m=+150.483826734" watchObservedRunningTime="2025-12-03 06:52:37.695787162 +0000 UTC m=+150.492477281" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.712117 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.713662 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.715987 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7nc8s"] Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.720253 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcb2j" podStartSLOduration=125.720231191 podStartE2EDuration="2m5.720231191s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.718495572 +0000 UTC m=+150.515185681" watchObservedRunningTime="2025-12-03 06:52:37.720231191 +0000 UTC m=+150.516921300" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.768810 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" podStartSLOduration=125.76879376 podStartE2EDuration="2m5.76879376s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:37.76703068 +0000 UTC m=+150.563720789" watchObservedRunningTime="2025-12-03 06:52:37.76879376 +0000 UTC m=+150.565483869" Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.787953 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.788904 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.288885975 +0000 UTC m=+151.085576084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.889221 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.889546 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.389531624 +0000 UTC m=+151.186221733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:37 crc kubenswrapper[4946]: I1203 06:52:37.993897 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:37 crc kubenswrapper[4946]: E1203 06:52:37.994152 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.494139307 +0000 UTC m=+151.290829416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.054822 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.056671 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.096834 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.097260 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.097536 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.097824 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.597811212 +0000 UTC m=+151.394501321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.114372 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.118595 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fsx6z"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.128894 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.162853 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.170925 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.198420 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.199637 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.699611834 +0000 UTC m=+151.496301933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.204308 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-55ddd"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.211013 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:38 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:38 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:38 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.211072 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.229341 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.246813 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.248803 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.259547 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzrcc"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.274177 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.309492 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.310093 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.310444 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.810432954 +0000 UTC m=+151.607123063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.338933 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-2q2qg"] Dec 03 06:52:38 crc kubenswrapper[4946]: W1203 06:52:38.341469 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0acb31c7_623c_40b7_889c_4a9a45b55cf1.slice/crio-a10f958c77ca61784d2004981f734334e9512ac4b97043cb13ab880cd5bc852c WatchSource:0}: Error finding container a10f958c77ca61784d2004981f734334e9512ac4b97043cb13ab880cd5bc852c: Status 404 returned error can't find the container with id a10f958c77ca61784d2004981f734334e9512ac4b97043cb13ab880cd5bc852c Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.355456 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7sng8"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.388073 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qsg2m"] Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.413149 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.413501 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:38.913485272 +0000 UTC m=+151.710175381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.514993 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.515615 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.015599823 +0000 UTC m=+151.812289932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.617793 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.618098 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.118083485 +0000 UTC m=+151.914773594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.661064 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" event={"ID":"670f4f60-09f3-4a59-bc86-b34fe9466e2e","Type":"ContainerStarted","Data":"aebfb118597562589392d091fe4377a0e86731a803ce0811fdf49d3dc67d66b5"} Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.665756 4946 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1bd7b04_110e_45fb_b972_c662a2e7c791.slice/crio-59da07418ee31534595967420e2ff74ed41ebf17a48a4a0967859198b6c31274.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.700968 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" event={"ID":"f4bf5842-4786-458d-9cdd-8f77b5bbce1f","Type":"ContainerStarted","Data":"ab6c0e55bc1e1db8551cb48c8df2010f1fd0acba58f2536b8e2c9b39f261a448"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.702808 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" podStartSLOduration=126.702790698 podStartE2EDuration="2m6.702790698s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:38.701128551 +0000 UTC m=+151.497818670" watchObservedRunningTime="2025-12-03 06:52:38.702790698 +0000 UTC m=+151.499480807" Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.714856 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" event={"ID":"8532cd3f-9a90-42cd-b237-44e18b9117aa","Type":"ContainerStarted","Data":"08b830a25c0b2a07c46858dcec4a3bf42553d621bbd810d021e44864f4520fb0"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.722868 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.723208 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.223194982 +0000 UTC m=+152.019885091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.759160 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" event={"ID":"a3aa4249-2022-45e7-844a-25e6b14f0661","Type":"ContainerStarted","Data":"9f984c0e5e02789ee5344512d8053a473970893fd39d7a2212db92d0c36e0dd5"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.775339 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" event={"ID":"9a05ab86-ca6d-418d-a3a7-cd8f958bca06","Type":"ContainerStarted","Data":"89257e29a31134a3359e9885be7d782a6d02207e4a18c7444500fb71897e38b7"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.777207 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.790929 4946 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-c4xq7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.791434 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" podUID="9a05ab86-ca6d-418d-a3a7-cd8f958bca06" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.794940 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qsg2m" event={"ID":"cbc2e507-308f-4be7-af6c-c356d3e7d2c5","Type":"ContainerStarted","Data":"46f7f8e451921f981344518fae16482ea65ee4b2865048f091bea2da1c44ab5f"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.798878 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" event={"ID":"554d7c42-1f14-4bac-b062-bdb816f18789","Type":"ContainerStarted","Data":"f5a58c4790e9e6272dade94af85a9fb30498ebf65d13572b68b16642e3766849"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.825088 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.826669 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.326651281 +0000 UTC m=+152.123341390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.856158 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" event={"ID":"cafb5284-e531-4979-86aa-ff4c626b71be","Type":"ContainerStarted","Data":"3264eb134c76072fee4e284283f7a0afbc9862e0654f6dcbd9b686ab1cd01938"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.856228 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" event={"ID":"cafb5284-e531-4979-86aa-ff4c626b71be","Type":"ContainerStarted","Data":"4a1f1d31350bf928344e6ba182a51b398e64c1fda544ef09dc1033fcfb7c246f"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.872576 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" event={"ID":"1b905b7b-c2ff-4cf1-9102-d7124d618ec7","Type":"ContainerStarted","Data":"1d20244f1d7f125c01da715ef7b06627769b0cda1627122b0f37be9eacc139df"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.886644 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" event={"ID":"0acb31c7-623c-40b7-889c-4a9a45b55cf1","Type":"ContainerStarted","Data":"a10f958c77ca61784d2004981f734334e9512ac4b97043cb13ab880cd5bc852c"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.891780 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" podStartSLOduration=125.891765614 podStartE2EDuration="2m5.891765614s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:38.817408837 +0000 UTC m=+151.614098946" watchObservedRunningTime="2025-12-03 06:52:38.891765614 +0000 UTC m=+151.688455723" Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.915358 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" event={"ID":"f27b8b97-2934-4d16-a756-8d402ec86bad","Type":"ContainerStarted","Data":"9c1710cdda6e7f2ebee69c8eec02ee85a90fab899d76939b1c77ad1efd49230e"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.915593 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" event={"ID":"f27b8b97-2934-4d16-a756-8d402ec86bad","Type":"ContainerStarted","Data":"9420d9a3f0d88ac47bb46537d73787f4fdb1886519c6caaff5c88e446a37cbb4"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.915676 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" event={"ID":"f27b8b97-2934-4d16-a756-8d402ec86bad","Type":"ContainerStarted","Data":"ac28746c7d42a51a6a68cd47b215a7c7ab1ee402ad66c5b26e91d2152c3ae291"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.927708 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:38 crc kubenswrapper[4946]: E1203 06:52:38.929448 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.429436272 +0000 UTC m=+152.226126371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.937042 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" event={"ID":"4a4c7b48-79a1-4a55-9012-a50c96908e7a","Type":"ContainerStarted","Data":"fc3831aa3c77bd212bae9a6f1afd4585d101171a2b312c6764da380c58fd5fcc"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.937180 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" event={"ID":"4a4c7b48-79a1-4a55-9012-a50c96908e7a","Type":"ContainerStarted","Data":"37eed05599a02655a0272ea0c431812c7a70df595ac952aeef165882c7901f51"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.951650 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-phv7g" podStartSLOduration=126.951631507 podStartE2EDuration="2m6.951631507s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:38.947584421 +0000 UTC m=+151.744274530" watchObservedRunningTime="2025-12-03 06:52:38.951631507 +0000 UTC m=+151.748321616" Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.952584 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qwfgf" podStartSLOduration=126.952576954 podStartE2EDuration="2m6.952576954s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:38.893218986 +0000 UTC m=+151.689909095" watchObservedRunningTime="2025-12-03 06:52:38.952576954 +0000 UTC m=+151.749267063" Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.967180 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" event={"ID":"ece5ff47-76df-4fad-9a59-3dbba27b788d","Type":"ContainerStarted","Data":"9fe5a496e319b9b388b37e12133d2357b1c3c1f53eeccd34542de231b00f6e6f"} Dec 03 06:52:38 crc kubenswrapper[4946]: I1203 06:52:38.982764 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" event={"ID":"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599","Type":"ContainerStarted","Data":"4f06cb00f8ed169138c0819478b881c96f8cdc33502d019f362849173df89d1f"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.018928 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w6x6n" event={"ID":"b43afb98-1ebd-4df2-85ea-c3c12c66ded1","Type":"ContainerStarted","Data":"b800830c3727a0c67ff9e49f0c678bb76869348eba19ce867879cc0a9c65aaa4"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.019774 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.032466 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.033665 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.533648782 +0000 UTC m=+152.330338891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.052239 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ghqsn" podStartSLOduration=127.052222193 podStartE2EDuration="2m7.052222193s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:38.974437779 +0000 UTC m=+151.771127888" watchObservedRunningTime="2025-12-03 06:52:39.052222193 +0000 UTC m=+151.848912302" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.104169 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" event={"ID":"7ec05e39-71b1-46b8-a6e4-8312e4d6d020","Type":"ContainerStarted","Data":"353f487220006a6a194c1d7416962dd18476c653e931da5863813227bd546c5a"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.132082 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gfz9s" podStartSLOduration=127.132062077 podStartE2EDuration="2m7.132062077s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:39.051096361 +0000 UTC m=+151.847786470" watchObservedRunningTime="2025-12-03 06:52:39.132062077 +0000 UTC m=+151.928752186" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.132439 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-w6x6n" podStartSLOduration=127.132433578 podStartE2EDuration="2m7.132433578s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:39.129415601 +0000 UTC m=+151.926105710" watchObservedRunningTime="2025-12-03 06:52:39.132433578 +0000 UTC m=+151.929123687" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.137135 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.137365 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-55ddd" event={"ID":"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9","Type":"ContainerStarted","Data":"7e20170973596f2990f93d58f72afcd27a3fb57fb489f98168c6674917aa2e31"} Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.138490 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.63846986 +0000 UTC m=+152.435159969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.152608 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" event={"ID":"1547b1af-d1e0-44a9-9831-30c901233123","Type":"ContainerStarted","Data":"00e7385f83cfd3d292b3449c85e6ea087630addfe670805101bad9bed74da1dd"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.154950 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2q2qg" event={"ID":"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7","Type":"ContainerStarted","Data":"89db323d9ed94b006f80128503a73475170e7aa2fdac4af6fa954084cc9395b9"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.173725 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" event={"ID":"c58dec3d-f028-43b2-8e71-1954da44b01c","Type":"ContainerStarted","Data":"3937ca50bbbcbd985b03b38dc3e5582bca9f67be52b48b7aa7f0a91ece61ed6a"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.215554 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" event={"ID":"7ef124c8-9694-4320-bdda-2ecc10489225","Type":"ContainerStarted","Data":"93311b066ec6232aa01a7fee5b8cc6e657bdb7c65796ffc83b971d0a3748eb64"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.226988 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:39 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:39 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:39 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.227043 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.239612 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" event={"ID":"f1bd7b04-110e-45fb-b972-c662a2e7c791","Type":"ContainerStarted","Data":"b29f687a965f8f65b01d0da982f9fd69518a690c468dc3a0a1ebcaf4c7b9b227"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.241106 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.241655 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.741640942 +0000 UTC m=+152.538331051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.242336 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" event={"ID":"8afa818f-4048-49d9-a54a-3ba48f3de8cd","Type":"ContainerStarted","Data":"51721c6a06ff109c81496d2684f4d90712bff04536a2740452afc2faefbf5d07"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.249096 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" event={"ID":"09b65f3a-96c6-4342-802d-8f83502aee5d","Type":"ContainerStarted","Data":"f6efac5c2c1101d6dd2824b9da381ba5f921ab658ef3b00170e3f74e5a9d0fbe"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.280336 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" event={"ID":"33e1bf8e-409b-4546-9eaa-c6099871e3d2","Type":"ContainerStarted","Data":"6c167f876d2cabc420591329c816e4521b0af03bb44ab92ee32e01ebe466b24e"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.296120 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" event={"ID":"4254dd3d-9a3f-475a-855b-95f7101c7ab9","Type":"ContainerStarted","Data":"dd9960419f918782019b9590def085727184940c3b5dab556fea22d7b7086ee9"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.323644 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" event={"ID":"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b","Type":"ContainerStarted","Data":"975ad01a2c0d1bbb761769b518e3aa8cdf5716fbe71bafaa4505ca2711b599e6"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.326730 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" event={"ID":"e87e9694-6803-4745-8a15-4065e6722e08","Type":"ContainerStarted","Data":"8dcc250bfd3b053146a31275431ca1bb62923846f8364b8d2dedfd6579728456"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.326780 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" event={"ID":"e87e9694-6803-4745-8a15-4065e6722e08","Type":"ContainerStarted","Data":"dae7c6eae0bad2d0bc14254c34d24df1f54f8ff468592fbcf7b416c4c2bf9c38"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.332203 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" event={"ID":"84f01e09-8152-4a65-b71a-5456ea557a77","Type":"ContainerStarted","Data":"af65a9faacdd1e60d128b292b3bb4ce736f0e5943fd192337f7949fb9254cd0a"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.332236 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.343502 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.344312 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.844297188 +0000 UTC m=+152.640987297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.354592 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" event={"ID":"6050037f-7b92-4a14-8597-3b6be013e8c9","Type":"ContainerStarted","Data":"430a56db4a4eb66b403b1dd81b83a721d76f469c114e12d211a072ad9d5aab60"} Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.357360 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.358908 4946 patch_prober.go:28] interesting pod/downloads-7954f5f757-ncr4w container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.358977 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ncr4w" podUID="ca8d6db9-e160-4ad7-a399-3dc1618add0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.406052 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.412451 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-w6x6n" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.446087 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.447469 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:39.947448749 +0000 UTC m=+152.744138858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.483461 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" podStartSLOduration=127.483436848 podStartE2EDuration="2m7.483436848s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:39.39014097 +0000 UTC m=+152.186831079" watchObservedRunningTime="2025-12-03 06:52:39.483436848 +0000 UTC m=+152.280126957" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.523942 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" podStartSLOduration=127.523927107 podStartE2EDuration="2m7.523927107s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:39.484808788 +0000 UTC m=+152.281498897" watchObservedRunningTime="2025-12-03 06:52:39.523927107 +0000 UTC m=+152.320617216" Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.554813 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.555243 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.055227782 +0000 UTC m=+152.851917881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.657855 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.658416 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.158400384 +0000 UTC m=+152.955090493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.770899 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.771907 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.27188932 +0000 UTC m=+153.068579429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.874267 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.874635 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.374621039 +0000 UTC m=+153.171311148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:39 crc kubenswrapper[4946]: I1203 06:52:39.976546 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:39 crc kubenswrapper[4946]: E1203 06:52:39.977570 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.477552763 +0000 UTC m=+153.274242882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.084984 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.085373 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.585350737 +0000 UTC m=+153.382040846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.188132 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.188494 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.688478887 +0000 UTC m=+153.485168996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.202969 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:40 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:40 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:40 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.203035 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.293414 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.293939 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.793918023 +0000 UTC m=+153.590608132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.404099 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.405375 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:40.905355541 +0000 UTC m=+153.702045650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.425565 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-drm7q" event={"ID":"e87e9694-6803-4745-8a15-4065e6722e08","Type":"ContainerStarted","Data":"3040c4da52f90ad757e404a833a01c2a583a281067ed81fa0a0fef1a4cca9c1a"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.439193 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" event={"ID":"8532cd3f-9a90-42cd-b237-44e18b9117aa","Type":"ContainerStarted","Data":"f84f20f3e39aea557645c0f4cc2d4524405b3fefe82f7b01eaf6ceebd2816dbe"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.458813 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gzfs9" podStartSLOduration=128.458785819 podStartE2EDuration="2m8.458785819s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:40.458291245 +0000 UTC m=+153.254981354" watchObservedRunningTime="2025-12-03 06:52:40.458785819 +0000 UTC m=+153.255475928" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.460975 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" event={"ID":"7ef124c8-9694-4320-bdda-2ecc10489225","Type":"ContainerStarted","Data":"f52c7024e3ad79ce94cc075d92931aa3cb19dda9a8cd7a671fdac6f92d986e15"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.476014 4946 generic.go:334] "Generic (PLEG): container finished" podID="4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599" containerID="e8803df46662d9b51c18d9899cf3b146d5c4523eb8dd1e25d65efe8d68c35240" exitCode=0 Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.476549 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" event={"ID":"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599","Type":"ContainerDied","Data":"e8803df46662d9b51c18d9899cf3b146d5c4523eb8dd1e25d65efe8d68c35240"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.482401 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" event={"ID":"8afa818f-4048-49d9-a54a-3ba48f3de8cd","Type":"ContainerStarted","Data":"30db0fc96e389924d8bef4fe4d3252f5ab721a1e4741709bd2826d02297989ea"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.498982 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" event={"ID":"0acb31c7-623c-40b7-889c-4a9a45b55cf1","Type":"ContainerStarted","Data":"cd212e0da299f93ef180a15e925ba5efb1009a1c072e28c32ac0913bb6d712c5"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.506367 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.506770 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.006735561 +0000 UTC m=+153.803425660 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.509396 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" event={"ID":"33e1bf8e-409b-4546-9eaa-c6099871e3d2","Type":"ContainerStarted","Data":"80d3d43b680f309770a94dd7ef84de61409f4c21a0cbf5a7b84ecbd76b17239c"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.525802 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-nt44m" podStartSLOduration=128.525779286 podStartE2EDuration="2m8.525779286s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:40.525327523 +0000 UTC m=+153.322017632" watchObservedRunningTime="2025-12-03 06:52:40.525779286 +0000 UTC m=+153.322469395" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.576305 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" event={"ID":"1b905b7b-c2ff-4cf1-9102-d7124d618ec7","Type":"ContainerStarted","Data":"d2646a5d633e47d41cfa3926cd68cd9c9dbe8a97099ab1c6161407206787b468"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.588030 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b926l" podStartSLOduration=128.588004656 podStartE2EDuration="2m8.588004656s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:40.575034685 +0000 UTC m=+153.371724794" watchObservedRunningTime="2025-12-03 06:52:40.588004656 +0000 UTC m=+153.384694775" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.623199 4946 generic.go:334] "Generic (PLEG): container finished" podID="f1bd7b04-110e-45fb-b972-c662a2e7c791" containerID="59da07418ee31534595967420e2ff74ed41ebf17a48a4a0967859198b6c31274" exitCode=0 Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.623296 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" event={"ID":"f1bd7b04-110e-45fb-b972-c662a2e7c791","Type":"ContainerDied","Data":"59da07418ee31534595967420e2ff74ed41ebf17a48a4a0967859198b6c31274"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.626860 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.629763 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.129712919 +0000 UTC m=+153.926403018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.698098 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" event={"ID":"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b","Type":"ContainerStarted","Data":"355012af05ffc202a7e2a63c223304b3c97e1e698e7b6a58e2666a3d8fd8ae5c"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.701116 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.726122 4946 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gzrcc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.726179 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.728244 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.728639 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.228623348 +0000 UTC m=+154.025313457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.764869 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" event={"ID":"1547b1af-d1e0-44a9-9831-30c901233123","Type":"ContainerStarted","Data":"fef680358fe87d85932256c515de17fec6fe6b68f9addae833bee2cfedb7628f"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.765605 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.789633 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-7sng8" podStartSLOduration=127.789607363 podStartE2EDuration="2m7.789607363s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:40.78776806 +0000 UTC m=+153.584458179" watchObservedRunningTime="2025-12-03 06:52:40.789607363 +0000 UTC m=+153.586297462" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.789980 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5trw" podStartSLOduration=127.789972243 podStartE2EDuration="2m7.789972243s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:40.756427464 +0000 UTC m=+153.553117573" watchObservedRunningTime="2025-12-03 06:52:40.789972243 +0000 UTC m=+153.586662342" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.830498 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.832178 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.33216259 +0000 UTC m=+154.128852699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.843429 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" event={"ID":"7ec05e39-71b1-46b8-a6e4-8312e4d6d020","Type":"ContainerStarted","Data":"5ed168f7ef9dc7e27bb987e704d5fb9fb4835ba15630c1c773a398f3470619b9"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.870204 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qsg2m" event={"ID":"cbc2e507-308f-4be7-af6c-c356d3e7d2c5","Type":"ContainerStarted","Data":"9a41843a688e5d4c3207e8c8a0e1d8697c4f7f51f5adac8b0649ea7857e2f9f4"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.896506 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" event={"ID":"09b65f3a-96c6-4342-802d-8f83502aee5d","Type":"ContainerStarted","Data":"c373702eea0e88bf0a2b60753206c63e4b510d384f2a0e39bfeb3093518b344b"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.908398 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" event={"ID":"554d7c42-1f14-4bac-b062-bdb816f18789","Type":"ContainerStarted","Data":"f366f87862202951eb0c120064bec4ca614f2021db3656e108f4d53e6d305e8d"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.931527 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" podStartSLOduration=127.931507662 podStartE2EDuration="2m7.931507662s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:40.85243075 +0000 UTC m=+153.649120859" watchObservedRunningTime="2025-12-03 06:52:40.931507662 +0000 UTC m=+153.728197771" Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.931782 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:40 crc kubenswrapper[4946]: E1203 06:52:40.933432 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.433416907 +0000 UTC m=+154.230107016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.946006 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-698gl" event={"ID":"670f4f60-09f3-4a59-bc86-b34fe9466e2e","Type":"ContainerStarted","Data":"83588dff76aecae72fa8dc3d4eebb59ab63078a77df6edc3d9730a4157c92b18"} Dec 03 06:52:40 crc kubenswrapper[4946]: I1203 06:52:40.978088 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" podStartSLOduration=127.978065574 podStartE2EDuration="2m7.978065574s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:40.93038057 +0000 UTC m=+153.727070679" watchObservedRunningTime="2025-12-03 06:52:40.978065574 +0000 UTC m=+153.774755683" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.008105 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" event={"ID":"a3aa4249-2022-45e7-844a-25e6b14f0661","Type":"ContainerStarted","Data":"24413e9ea0e1effa1db7990493d92e64a01971093916239d24301e83ba803420"} Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.008178 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" event={"ID":"a3aa4249-2022-45e7-844a-25e6b14f0661","Type":"ContainerStarted","Data":"9b9bb054ee0c9f397f4f7ecd73fcdbb9283e9e9932877cb7fe45855a5e0e8166"} Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.010864 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" event={"ID":"9a05ab86-ca6d-418d-a3a7-cd8f958bca06","Type":"ContainerStarted","Data":"c3000dc3971473402ab1f1c40c4340ae1b046a60ae3d15dcd731187842893014"} Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.021022 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" event={"ID":"f4bf5842-4786-458d-9cdd-8f77b5bbce1f","Type":"ContainerStarted","Data":"5a10295e32a72b28a8d05ecc3f84924bd643d3991a2fa62df3ac6c77bcfb34c2"} Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.022036 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.036838 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" event={"ID":"4254dd3d-9a3f-475a-855b-95f7101c7ab9","Type":"ContainerStarted","Data":"5ce20789cbd7006f2fbfd6aa043cbac52e0e39c2e7d317f6e1f66ff2b8dfb85b"} Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.037689 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.045670 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.046178 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.546157392 +0000 UTC m=+154.342847501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.058510 4946 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fgbb8 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.058590 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" podUID="4254dd3d-9a3f-475a-855b-95f7101c7ab9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.063066 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" event={"ID":"c58dec3d-f028-43b2-8e71-1954da44b01c","Type":"ContainerStarted","Data":"94f485723b01d8266eef0172918fa3052a8d3ef7cfb2c59b2025e56ad061883d"} Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.063174 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f59l8" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.063722 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.124016 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wr8vk" podStartSLOduration=128.123988028 podStartE2EDuration="2m8.123988028s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:40.984295972 +0000 UTC m=+153.780986081" watchObservedRunningTime="2025-12-03 06:52:41.123988028 +0000 UTC m=+153.920678157" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.129811 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qsg2m" podStartSLOduration=9.129800584 podStartE2EDuration="9.129800584s" podCreationTimestamp="2025-12-03 06:52:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:41.121435015 +0000 UTC m=+153.918125134" watchObservedRunningTime="2025-12-03 06:52:41.129800584 +0000 UTC m=+153.926490693" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.146955 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.148781 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.648762467 +0000 UTC m=+154.445452576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.200399 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:41 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:41 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:41 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.200440 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.249863 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.250849 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.750800406 +0000 UTC m=+154.547490515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.298790 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9spfp" podStartSLOduration=129.298772248 podStartE2EDuration="2m9.298772248s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:41.199627262 +0000 UTC m=+153.996317371" watchObservedRunningTime="2025-12-03 06:52:41.298772248 +0000 UTC m=+154.095462357" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.300626 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" podStartSLOduration=128.300621791 podStartE2EDuration="2m8.300621791s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:41.298102869 +0000 UTC m=+154.094792988" watchObservedRunningTime="2025-12-03 06:52:41.300621791 +0000 UTC m=+154.097311900" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.352375 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.352716 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.852702841 +0000 UTC m=+154.649392950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.457856 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.458356 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:41.958340483 +0000 UTC m=+154.755030592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.491518 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bq9b" podStartSLOduration=129.491503601 podStartE2EDuration="2m9.491503601s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:41.378995803 +0000 UTC m=+154.175685912" watchObservedRunningTime="2025-12-03 06:52:41.491503601 +0000 UTC m=+154.288193710" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.555574 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" podStartSLOduration=128.555551813 podStartE2EDuration="2m8.555551813s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:41.555109261 +0000 UTC m=+154.351799370" watchObservedRunningTime="2025-12-03 06:52:41.555551813 +0000 UTC m=+154.352241922" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.559211 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.559405 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.059348372 +0000 UTC m=+154.856038481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.559444 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.559952 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.059943089 +0000 UTC m=+154.856633198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.602064 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xtwj9" podStartSLOduration=128.602041763 podStartE2EDuration="2m8.602041763s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:41.599573723 +0000 UTC m=+154.396263832" watchObservedRunningTime="2025-12-03 06:52:41.602041763 +0000 UTC m=+154.398731872" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.661187 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.661914 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.161893035 +0000 UTC m=+154.958583134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.711937 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c4xq7" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.744065 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" podStartSLOduration=129.744031815 podStartE2EDuration="2m9.744031815s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:41.696915827 +0000 UTC m=+154.493605936" watchObservedRunningTime="2025-12-03 06:52:41.744031815 +0000 UTC m=+154.540721924" Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.770436 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.771115 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.271101989 +0000 UTC m=+155.067792098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.873220 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.874204 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.374184988 +0000 UTC m=+155.170875097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:41 crc kubenswrapper[4946]: I1203 06:52:41.975203 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:41 crc kubenswrapper[4946]: E1203 06:52:41.975941 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.475924809 +0000 UTC m=+155.272614918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.071815 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-55ddd" event={"ID":"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9","Type":"ContainerStarted","Data":"cc7429f3e96f17d6fb85e98c5f39f2528100db4a3dacbeeeffb304faee04cd51"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.073410 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" event={"ID":"1547b1af-d1e0-44a9-9831-30c901233123","Type":"ContainerStarted","Data":"a9ea9909f29af74341f62779043c5e751c0dc535a33f0703e593dcf9637a67a3"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.075052 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" event={"ID":"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599","Type":"ContainerStarted","Data":"8a06f718d68142f4d60c24a9a14d639d73c5a23ffbaf5feb6961583462b5e9c3"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.075563 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.075825 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.575795945 +0000 UTC m=+155.372486054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.075854 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.076191 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.576176776 +0000 UTC m=+155.372866885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.076541 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2q2qg" event={"ID":"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7","Type":"ContainerStarted","Data":"9d1dc72838aaa9f31b5ed538aaa2973efb298dbc3db3bb25e240721d9b3fc824"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.076616 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2q2qg" event={"ID":"bacf1b15-c2aa-4d79-8bc9-2977deea4ef7","Type":"ContainerStarted","Data":"f2660243b8b32a808d3c4585fe8010087cc61cf06f1d6fcc579606dfc4634acb"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.076868 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.078338 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" event={"ID":"6050037f-7b92-4a14-8597-3b6be013e8c9","Type":"ContainerStarted","Data":"69fc57b251605fc89141e8bd2c74e02d457116db228e735120479d8da7f9a624"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.080390 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9927" event={"ID":"09b65f3a-96c6-4342-802d-8f83502aee5d","Type":"ContainerStarted","Data":"6f5c77fa12f7f009e0c3b02283f7548402bd79ca16ee54d5a23534dada562924"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.083820 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" event={"ID":"1b905b7b-c2ff-4cf1-9102-d7124d618ec7","Type":"ContainerStarted","Data":"f73a2e958935621dc722ad38bd078670ff071aa8fb4e11c4887b59ae4bab4ec3"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.087472 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" event={"ID":"f1bd7b04-110e-45fb-b972-c662a2e7c791","Type":"ContainerStarted","Data":"12311661392f9ccadef7760a8f20c6afeb557335c8cbbd78cc80164038b7cbb5"} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.088645 4946 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gzrcc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.088714 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.101839 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-2q2qg" podStartSLOduration=10.101800069 podStartE2EDuration="10.101800069s" podCreationTimestamp="2025-12-03 06:52:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:42.101337506 +0000 UTC m=+154.898027635" watchObservedRunningTime="2025-12-03 06:52:42.101800069 +0000 UTC m=+154.898490178" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.157637 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" podStartSLOduration=129.157586235 podStartE2EDuration="2m9.157586235s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:42.150895494 +0000 UTC m=+154.947585613" watchObservedRunningTime="2025-12-03 06:52:42.157586235 +0000 UTC m=+154.954276684" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.160902 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fgbb8" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.162477 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f6b24" podStartSLOduration=130.162466555 podStartE2EDuration="2m10.162466555s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:42.128881754 +0000 UTC m=+154.925571863" watchObservedRunningTime="2025-12-03 06:52:42.162466555 +0000 UTC m=+154.959156664" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.178373 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.180489 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.68046454 +0000 UTC m=+155.477154649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.195224 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:42 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:42 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:42 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.195264 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.196178 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fsx6z" podStartSLOduration=129.196165699 podStartE2EDuration="2m9.196165699s" podCreationTimestamp="2025-12-03 06:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:42.181848369 +0000 UTC m=+154.978538488" watchObservedRunningTime="2025-12-03 06:52:42.196165699 +0000 UTC m=+154.992855808" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.284987 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.285521 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.785500114 +0000 UTC m=+155.582190223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.386438 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.386697 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.886655448 +0000 UTC m=+155.683345557 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.387289 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.387810 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.887800131 +0000 UTC m=+155.684490240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.494093 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.494371 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.994332398 +0000 UTC m=+155.791022507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.494461 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.494865 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:42.994846503 +0000 UTC m=+155.791536612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.595790 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.596083 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:43.096043717 +0000 UTC m=+155.892733826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.596395 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.596864 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:43.09685081 +0000 UTC m=+155.893540919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.696153 4946 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.697869 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.698044 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:43.198012654 +0000 UTC m=+155.994702763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.698177 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.698514 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:52:43.198507598 +0000 UTC m=+155.995197707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-42855" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.799184 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:42 crc kubenswrapper[4946]: E1203 06:52:42.799521 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:52:43.299507257 +0000 UTC m=+156.096197366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.830931 4946 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T06:52:42.696184091Z","Handler":null,"Name":""} Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.834074 4946 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.834137 4946 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.901470 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.903922 4946 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.903968 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:42 crc kubenswrapper[4946]: I1203 06:52:42.960898 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-42855\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.004043 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.035852 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hv9ql"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.036914 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.038887 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.047690 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.051493 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hv9ql"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.093670 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" event={"ID":"4ea9895d-fdfd-4cc4-a8cb-2b9fe0905599","Type":"ContainerStarted","Data":"5c103a33ed51e0ba0653770174397184a0ce1b4e1953e759d0c3afc21ef54385"} Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.097415 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-55ddd" event={"ID":"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9","Type":"ContainerStarted","Data":"76cc77bfe01dca7e4dd3133795c8a77aa8ec6d1f3fe3d083420f01ffce932efd"} Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.097464 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-55ddd" event={"ID":"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9","Type":"ContainerStarted","Data":"99556e83d9c7805c4c572abf1ee5c68671ea3241290ba249612cf2ccb8bc3197"} Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.097505 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-55ddd" event={"ID":"f91ad74f-43c3-4c3e-b9a9-0e7021d2d4c9","Type":"ContainerStarted","Data":"648efe2172b7d280f1909ffc776041640a3ba74964e8f465a690255efae728a8"} Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.099627 4946 generic.go:334] "Generic (PLEG): container finished" podID="c58dec3d-f028-43b2-8e71-1954da44b01c" containerID="94f485723b01d8266eef0172918fa3052a8d3ef7cfb2c59b2025e56ad061883d" exitCode=0 Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.100304 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" event={"ID":"c58dec3d-f028-43b2-8e71-1954da44b01c","Type":"ContainerDied","Data":"94f485723b01d8266eef0172918fa3052a8d3ef7cfb2c59b2025e56ad061883d"} Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.105729 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwqbg\" (UniqueName: \"kubernetes.io/projected/0ce0d8cc-befc-48b1-b330-a132440b36a4-kube-api-access-gwqbg\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.105846 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-catalog-content\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.105875 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-utilities\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.118892 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" podStartSLOduration=131.118868493 podStartE2EDuration="2m11.118868493s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:43.118290186 +0000 UTC m=+155.914980305" watchObservedRunningTime="2025-12-03 06:52:43.118868493 +0000 UTC m=+155.915558602" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.155671 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-55ddd" podStartSLOduration=11.155646845 podStartE2EDuration="11.155646845s" podCreationTimestamp="2025-12-03 06:52:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:43.154840082 +0000 UTC m=+155.951530201" watchObservedRunningTime="2025-12-03 06:52:43.155646845 +0000 UTC m=+155.952336954" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.194607 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:43 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:43 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:43 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.194701 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.207251 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwqbg\" (UniqueName: \"kubernetes.io/projected/0ce0d8cc-befc-48b1-b330-a132440b36a4-kube-api-access-gwqbg\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.207427 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-catalog-content\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.207483 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-utilities\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.210415 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-catalog-content\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.210638 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-utilities\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.231358 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.231427 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zfzbf"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.232528 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.234772 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.241763 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwqbg\" (UniqueName: \"kubernetes.io/projected/0ce0d8cc-befc-48b1-b330-a132440b36a4-kube-api-access-gwqbg\") pod \"community-operators-hv9ql\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.257293 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zfzbf"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.309529 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mktw6\" (UniqueName: \"kubernetes.io/projected/a48e7204-9569-494d-b0bc-af6cc81f0f4a-kube-api-access-mktw6\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.309978 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-utilities\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.310010 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-catalog-content\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.354173 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.422658 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mktw6\" (UniqueName: \"kubernetes.io/projected/a48e7204-9569-494d-b0bc-af6cc81f0f4a-kube-api-access-mktw6\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.422772 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-utilities\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.422792 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-catalog-content\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.423546 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-catalog-content\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.424164 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-utilities\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.447886 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ch5kh"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.449408 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.458976 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mktw6\" (UniqueName: \"kubernetes.io/projected/a48e7204-9569-494d-b0bc-af6cc81f0f4a-kube-api-access-mktw6\") pod \"certified-operators-zfzbf\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.475805 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ch5kh"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.523673 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n45sl\" (UniqueName: \"kubernetes.io/projected/6f544f2a-7068-4378-a9f2-3ba25efd90a9-kube-api-access-n45sl\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.524572 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-utilities\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.524606 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-catalog-content\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.580264 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.604181 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.625991 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n45sl\" (UniqueName: \"kubernetes.io/projected/6f544f2a-7068-4378-a9f2-3ba25efd90a9-kube-api-access-n45sl\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.626390 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-utilities\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.626415 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-catalog-content\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.626983 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-catalog-content\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.627317 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-utilities\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.638809 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fm4jx"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.640986 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.649375 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n45sl\" (UniqueName: \"kubernetes.io/projected/6f544f2a-7068-4378-a9f2-3ba25efd90a9-kube-api-access-n45sl\") pod \"community-operators-ch5kh\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.673771 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fm4jx"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.727772 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-utilities\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.727850 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-catalog-content\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.727905 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jg9g\" (UniqueName: \"kubernetes.io/projected/60435668-f809-46b8-bea8-9f2a186aa15f-kube-api-access-2jg9g\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.794506 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.829618 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-utilities\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.829689 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-catalog-content\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.829750 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jg9g\" (UniqueName: \"kubernetes.io/projected/60435668-f809-46b8-bea8-9f2a186aa15f-kube-api-access-2jg9g\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.830600 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-utilities\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.830978 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-catalog-content\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.845921 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-42855"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.851204 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jg9g\" (UniqueName: \"kubernetes.io/projected/60435668-f809-46b8-bea8-9f2a186aa15f-kube-api-access-2jg9g\") pod \"certified-operators-fm4jx\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.851361 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hv9ql"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.887796 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zfzbf"] Dec 03 06:52:43 crc kubenswrapper[4946]: I1203 06:52:43.968641 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.112770 4946 generic.go:334] "Generic (PLEG): container finished" podID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerID="e2615c32a228cb97732278adaab0a48671250280d1c635c2b3b541a1b00afdf8" exitCode=0 Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.112870 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv9ql" event={"ID":"0ce0d8cc-befc-48b1-b330-a132440b36a4","Type":"ContainerDied","Data":"e2615c32a228cb97732278adaab0a48671250280d1c635c2b3b541a1b00afdf8"} Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.112929 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv9ql" event={"ID":"0ce0d8cc-befc-48b1-b330-a132440b36a4","Type":"ContainerStarted","Data":"14b7c7c2581e84048b2ecb3385aeb3c11c18262cc73d23b5a22a62b41a305587"} Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.116372 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.117249 4946 generic.go:334] "Generic (PLEG): container finished" podID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerID="f131cf049caee8df568e80e621b4ed80d56def78ed9e80c326f1cc9a95cdaebe" exitCode=0 Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.117319 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfzbf" event={"ID":"a48e7204-9569-494d-b0bc-af6cc81f0f4a","Type":"ContainerDied","Data":"f131cf049caee8df568e80e621b4ed80d56def78ed9e80c326f1cc9a95cdaebe"} Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.117349 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfzbf" event={"ID":"a48e7204-9569-494d-b0bc-af6cc81f0f4a","Type":"ContainerStarted","Data":"1d70843e6e1d2c52f5158265161d1cfd5f9818975d4239bfe4afc58e6b573587"} Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.122475 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-42855" event={"ID":"8b8c7afc-56fb-49ab-ae82-e60167809f15","Type":"ContainerStarted","Data":"e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a"} Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.122553 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-42855" event={"ID":"8b8c7afc-56fb-49ab-ae82-e60167809f15","Type":"ContainerStarted","Data":"cd212aeffb2907e8043514a3141d5e7e8ec9e03ce3cba83d8ef1a446dcf78cd3"} Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.187497 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-42855" podStartSLOduration=132.187473821 podStartE2EDuration="2m12.187473821s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:52:44.162206298 +0000 UTC m=+156.958896427" watchObservedRunningTime="2025-12-03 06:52:44.187473821 +0000 UTC m=+156.984163940" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.194647 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:44 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:44 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:44 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.194711 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.230898 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fm4jx"] Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.285529 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ch5kh"] Dec 03 06:52:44 crc kubenswrapper[4946]: W1203 06:52:44.291382 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f544f2a_7068_4378_a9f2_3ba25efd90a9.slice/crio-9a35011ba226cb087237002e4f836f63e5a9a64f05d9f748075d2c610fb98b94 WatchSource:0}: Error finding container 9a35011ba226cb087237002e4f836f63e5a9a64f05d9f748075d2c610fb98b94: Status 404 returned error can't find the container with id 9a35011ba226cb087237002e4f836f63e5a9a64f05d9f748075d2c610fb98b94 Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.379241 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.440109 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c58dec3d-f028-43b2-8e71-1954da44b01c-secret-volume\") pod \"c58dec3d-f028-43b2-8e71-1954da44b01c\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.440262 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c58dec3d-f028-43b2-8e71-1954da44b01c-config-volume\") pod \"c58dec3d-f028-43b2-8e71-1954da44b01c\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.440387 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngm49\" (UniqueName: \"kubernetes.io/projected/c58dec3d-f028-43b2-8e71-1954da44b01c-kube-api-access-ngm49\") pod \"c58dec3d-f028-43b2-8e71-1954da44b01c\" (UID: \"c58dec3d-f028-43b2-8e71-1954da44b01c\") " Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.442017 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c58dec3d-f028-43b2-8e71-1954da44b01c-config-volume" (OuterVolumeSpecName: "config-volume") pod "c58dec3d-f028-43b2-8e71-1954da44b01c" (UID: "c58dec3d-f028-43b2-8e71-1954da44b01c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.451063 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58dec3d-f028-43b2-8e71-1954da44b01c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c58dec3d-f028-43b2-8e71-1954da44b01c" (UID: "c58dec3d-f028-43b2-8e71-1954da44b01c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.451751 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58dec3d-f028-43b2-8e71-1954da44b01c-kube-api-access-ngm49" (OuterVolumeSpecName: "kube-api-access-ngm49") pod "c58dec3d-f028-43b2-8e71-1954da44b01c" (UID: "c58dec3d-f028-43b2-8e71-1954da44b01c"). InnerVolumeSpecName "kube-api-access-ngm49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.542347 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngm49\" (UniqueName: \"kubernetes.io/projected/c58dec3d-f028-43b2-8e71-1954da44b01c-kube-api-access-ngm49\") on node \"crc\" DevicePath \"\"" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.542393 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c58dec3d-f028-43b2-8e71-1954da44b01c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.542403 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c58dec3d-f028-43b2-8e71-1954da44b01c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.964810 4946 patch_prober.go:28] interesting pod/downloads-7954f5f757-ncr4w container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.964874 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ncr4w" podUID="ca8d6db9-e160-4ad7-a399-3dc1618add0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.965486 4946 patch_prober.go:28] interesting pod/downloads-7954f5f757-ncr4w container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.965560 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ncr4w" podUID="ca8d6db9-e160-4ad7-a399-3dc1618add0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.995183 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:44 crc kubenswrapper[4946]: I1203 06:52:44.995248 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.001419 4946 patch_prober.go:28] interesting pod/console-f9d7485db-zkld2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.001481 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zkld2" podUID="a89bae42-673b-4eb1-87d2-ef2dd919c4d4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.133495 4946 generic.go:334] "Generic (PLEG): container finished" podID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerID="890a47d7b4c19a3bf12a50c521a092deca15e9345f1a940677f86b4a6cef2587" exitCode=0 Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.133570 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch5kh" event={"ID":"6f544f2a-7068-4378-a9f2-3ba25efd90a9","Type":"ContainerDied","Data":"890a47d7b4c19a3bf12a50c521a092deca15e9345f1a940677f86b4a6cef2587"} Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.133608 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch5kh" event={"ID":"6f544f2a-7068-4378-a9f2-3ba25efd90a9","Type":"ContainerStarted","Data":"9a35011ba226cb087237002e4f836f63e5a9a64f05d9f748075d2c610fb98b94"} Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.140825 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.140852 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.156110 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" event={"ID":"c58dec3d-f028-43b2-8e71-1954da44b01c","Type":"ContainerDied","Data":"3937ca50bbbcbd985b03b38dc3e5582bca9f67be52b48b7aa7f0a91ece61ed6a"} Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.156175 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3937ca50bbbcbd985b03b38dc3e5582bca9f67be52b48b7aa7f0a91ece61ed6a" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.156368 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.158850 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.161209 4946 generic.go:334] "Generic (PLEG): container finished" podID="60435668-f809-46b8-bea8-9f2a186aa15f" containerID="aca86845be52987ffcd848eb411401424d22f2e9a3fa9399459c86f66e3d51e7" exitCode=0 Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.163288 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm4jx" event={"ID":"60435668-f809-46b8-bea8-9f2a186aa15f","Type":"ContainerDied","Data":"aca86845be52987ffcd848eb411401424d22f2e9a3fa9399459c86f66e3d51e7"} Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.163332 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm4jx" event={"ID":"60435668-f809-46b8-bea8-9f2a186aa15f","Type":"ContainerStarted","Data":"5b104fafca2ba3a19808310c17ef8a6baed66dd6fd0a662588cdbb5b6a1523b8"} Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.164132 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.170636 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-7nc8s" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.190882 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.193821 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:45 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:45 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:45 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.193868 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.232549 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.232585 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.253155 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bd7s4"] Dec 03 06:52:45 crc kubenswrapper[4946]: E1203 06:52:45.253364 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58dec3d-f028-43b2-8e71-1954da44b01c" containerName="collect-profiles" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.253376 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58dec3d-f028-43b2-8e71-1954da44b01c" containerName="collect-profiles" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.253468 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c58dec3d-f028-43b2-8e71-1954da44b01c" containerName="collect-profiles" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.256014 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.265959 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.272827 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.340700 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd7s4"] Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.369393 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-catalog-content\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.369495 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-utilities\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.369526 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dgzr\" (UniqueName: \"kubernetes.io/projected/9d975a9e-23bf-4eec-a768-0b962d4b1022-kube-api-access-5dgzr\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.471091 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-catalog-content\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.471145 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-utilities\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.471542 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-utilities\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.471162 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dgzr\" (UniqueName: \"kubernetes.io/projected/9d975a9e-23bf-4eec-a768-0b962d4b1022-kube-api-access-5dgzr\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.472074 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-catalog-content\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.489150 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dgzr\" (UniqueName: \"kubernetes.io/projected/9d975a9e-23bf-4eec-a768-0b962d4b1022-kube-api-access-5dgzr\") pod \"redhat-marketplace-bd7s4\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.496388 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.597031 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.645116 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zckq8"] Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.646230 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.658409 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zckq8"] Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.781773 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-utilities\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.781864 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m7lz\" (UniqueName: \"kubernetes.io/projected/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-kube-api-access-2m7lz\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.781889 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-catalog-content\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.883439 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m7lz\" (UniqueName: \"kubernetes.io/projected/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-kube-api-access-2m7lz\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.883515 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-catalog-content\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.884145 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-catalog-content\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.884211 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-utilities\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.884628 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-utilities\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.907557 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m7lz\" (UniqueName: \"kubernetes.io/projected/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-kube-api-access-2m7lz\") pod \"redhat-marketplace-zckq8\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:45 crc kubenswrapper[4946]: I1203 06:52:45.993842 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd7s4"] Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.018790 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.180037 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd7s4" event={"ID":"9d975a9e-23bf-4eec-a768-0b962d4b1022","Type":"ContainerStarted","Data":"9c041125aee77e945a3bf8d6c7ea38a71eb6c4cdddfbff257ddc84b46f5254af"} Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.189806 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7hmqc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.194290 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:46 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:46 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:46 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.194360 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.270401 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6bhbz"] Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.271726 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.275339 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.287879 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6bhbz"] Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.395407 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-catalog-content\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.395495 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gq22\" (UniqueName: \"kubernetes.io/projected/9e385355-ddac-4e63-935c-de1ad706600a-kube-api-access-6gq22\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.395528 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-utilities\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.505377 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gq22\" (UniqueName: \"kubernetes.io/projected/9e385355-ddac-4e63-935c-de1ad706600a-kube-api-access-6gq22\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.505847 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-utilities\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.506064 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-catalog-content\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.507046 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-catalog-content\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.507247 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-utilities\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.554714 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gq22\" (UniqueName: \"kubernetes.io/projected/9e385355-ddac-4e63-935c-de1ad706600a-kube-api-access-6gq22\") pod \"redhat-operators-6bhbz\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.603032 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.622111 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.623017 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.634838 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.635105 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.642673 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.645563 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bs2js"] Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.646707 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.670804 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bs2js"] Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.709460 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-utilities\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.709541 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-catalog-content\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.709580 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5add0485-955c-4aa2-8a84-6105e51f9360-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5add0485-955c-4aa2-8a84-6105e51f9360\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.709635 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtkls\" (UniqueName: \"kubernetes.io/projected/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-kube-api-access-wtkls\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.709717 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5add0485-955c-4aa2-8a84-6105e51f9360-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5add0485-955c-4aa2-8a84-6105e51f9360\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.762125 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zckq8"] Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.812491 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-catalog-content\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.812564 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5add0485-955c-4aa2-8a84-6105e51f9360-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5add0485-955c-4aa2-8a84-6105e51f9360\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.812594 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtkls\" (UniqueName: \"kubernetes.io/projected/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-kube-api-access-wtkls\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.812639 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5add0485-955c-4aa2-8a84-6105e51f9360-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5add0485-955c-4aa2-8a84-6105e51f9360\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.813040 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-utilities\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.813971 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-utilities\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.813970 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5add0485-955c-4aa2-8a84-6105e51f9360-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5add0485-955c-4aa2-8a84-6105e51f9360\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.814364 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-catalog-content\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.838417 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5add0485-955c-4aa2-8a84-6105e51f9360-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5add0485-955c-4aa2-8a84-6105e51f9360\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.845547 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtkls\" (UniqueName: \"kubernetes.io/projected/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-kube-api-access-wtkls\") pod \"redhat-operators-bs2js\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.972990 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.988113 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:52:46 crc kubenswrapper[4946]: I1203 06:52:46.998642 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6bhbz"] Dec 03 06:52:47 crc kubenswrapper[4946]: W1203 06:52:47.092877 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e385355_ddac_4e63_935c_de1ad706600a.slice/crio-e584718b3b0e661fad3a0fabc12582f5cc9024f3d7164cd58728af30de3e853d WatchSource:0}: Error finding container e584718b3b0e661fad3a0fabc12582f5cc9024f3d7164cd58728af30de3e853d: Status 404 returned error can't find the container with id e584718b3b0e661fad3a0fabc12582f5cc9024f3d7164cd58728af30de3e853d Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.190865 4946 generic.go:334] "Generic (PLEG): container finished" podID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerID="144f08daae3dcd0f31e0ac9e1c7bed3d2c3b9b6e7bd6de75c323aa483dc52c97" exitCode=0 Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.190943 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd7s4" event={"ID":"9d975a9e-23bf-4eec-a768-0b962d4b1022","Type":"ContainerDied","Data":"144f08daae3dcd0f31e0ac9e1c7bed3d2c3b9b6e7bd6de75c323aa483dc52c97"} Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.193585 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:47 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:47 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:47 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.193640 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.194332 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bhbz" event={"ID":"9e385355-ddac-4e63-935c-de1ad706600a","Type":"ContainerStarted","Data":"e584718b3b0e661fad3a0fabc12582f5cc9024f3d7164cd58728af30de3e853d"} Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.197018 4946 generic.go:334] "Generic (PLEG): container finished" podID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerID="05cd91c2746f63216e4e3ad8dbed14cf39060a8687381c43551f9bf5e8e5ad55" exitCode=0 Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.197113 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zckq8" event={"ID":"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c","Type":"ContainerDied","Data":"05cd91c2746f63216e4e3ad8dbed14cf39060a8687381c43551f9bf5e8e5ad55"} Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.197177 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zckq8" event={"ID":"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c","Type":"ContainerStarted","Data":"f46e52c05acb2e278cdafba8006fcf30c138e42fa92223a2e0fbf635654b9c3b"} Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.520689 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bs2js"] Dec 03 06:52:47 crc kubenswrapper[4946]: W1203 06:52:47.533474 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63a8d2b9_3245_443a_9c0c_dcd4f56a66a9.slice/crio-0c5bf7c24c764c2aa1c27bba67b01febcd796299c671df13453c202a37494e91 WatchSource:0}: Error finding container 0c5bf7c24c764c2aa1c27bba67b01febcd796299c671df13453c202a37494e91: Status 404 returned error can't find the container with id 0c5bf7c24c764c2aa1c27bba67b01febcd796299c671df13453c202a37494e91 Dec 03 06:52:47 crc kubenswrapper[4946]: I1203 06:52:47.578878 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 06:52:47 crc kubenswrapper[4946]: W1203 06:52:47.613849 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5add0485_955c_4aa2_8a84_6105e51f9360.slice/crio-6c1dba54eb478404b44652bcacbde25e4c23953f0a2a56a3f85160ab144a3f52 WatchSource:0}: Error finding container 6c1dba54eb478404b44652bcacbde25e4c23953f0a2a56a3f85160ab144a3f52: Status 404 returned error can't find the container with id 6c1dba54eb478404b44652bcacbde25e4c23953f0a2a56a3f85160ab144a3f52 Dec 03 06:52:48 crc kubenswrapper[4946]: I1203 06:52:48.196043 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:48 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:48 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:48 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:48 crc kubenswrapper[4946]: I1203 06:52:48.196561 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:48 crc kubenswrapper[4946]: I1203 06:52:48.209695 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bs2js" event={"ID":"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9","Type":"ContainerStarted","Data":"0c5bf7c24c764c2aa1c27bba67b01febcd796299c671df13453c202a37494e91"} Dec 03 06:52:48 crc kubenswrapper[4946]: I1203 06:52:48.223365 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5add0485-955c-4aa2-8a84-6105e51f9360","Type":"ContainerStarted","Data":"6c1dba54eb478404b44652bcacbde25e4c23953f0a2a56a3f85160ab144a3f52"} Dec 03 06:52:49 crc kubenswrapper[4946]: I1203 06:52:49.196041 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:49 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:49 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:49 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:49 crc kubenswrapper[4946]: I1203 06:52:49.196571 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:49 crc kubenswrapper[4946]: I1203 06:52:49.246535 4946 generic.go:334] "Generic (PLEG): container finished" podID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerID="62abc5d6c5e73ce76fd4ff84ff1a313c1f5dc611919072bebe188667ccd0b4f0" exitCode=0 Dec 03 06:52:49 crc kubenswrapper[4946]: I1203 06:52:49.246675 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bs2js" event={"ID":"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9","Type":"ContainerDied","Data":"62abc5d6c5e73ce76fd4ff84ff1a313c1f5dc611919072bebe188667ccd0b4f0"} Dec 03 06:52:49 crc kubenswrapper[4946]: I1203 06:52:49.251843 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5add0485-955c-4aa2-8a84-6105e51f9360","Type":"ContainerStarted","Data":"875619b8256236ee0a6759c80a8cf653ec1538d044f7df5fac156aed425ab2c1"} Dec 03 06:52:49 crc kubenswrapper[4946]: I1203 06:52:49.254854 4946 generic.go:334] "Generic (PLEG): container finished" podID="9e385355-ddac-4e63-935c-de1ad706600a" containerID="3fd32f9c91be3d53cb86a607657c9465796f34a1149dc4934cf25f29b96e82bb" exitCode=0 Dec 03 06:52:49 crc kubenswrapper[4946]: I1203 06:52:49.254901 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bhbz" event={"ID":"9e385355-ddac-4e63-935c-de1ad706600a","Type":"ContainerDied","Data":"3fd32f9c91be3d53cb86a607657c9465796f34a1149dc4934cf25f29b96e82bb"} Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.047894 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.053682 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.067921 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.069266 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.073475 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.194211 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.194288 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.195940 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:50 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:50 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:50 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.196015 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.263093 4946 generic.go:334] "Generic (PLEG): container finished" podID="5add0485-955c-4aa2-8a84-6105e51f9360" containerID="875619b8256236ee0a6759c80a8cf653ec1538d044f7df5fac156aed425ab2c1" exitCode=0 Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.263136 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5add0485-955c-4aa2-8a84-6105e51f9360","Type":"ContainerDied","Data":"875619b8256236ee0a6759c80a8cf653ec1538d044f7df5fac156aed425ab2c1"} Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.296058 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.296189 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.296272 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.317627 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.385723 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.730617 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 06:52:50 crc kubenswrapper[4946]: W1203 06:52:50.734693 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7b584fa1_0942_4df1_98e3_c6c4e1429c85.slice/crio-3df37b90178a71c6b9e5d94a067f0299b4d205eaf8837d34a1a73f04c03cbde9 WatchSource:0}: Error finding container 3df37b90178a71c6b9e5d94a067f0299b4d205eaf8837d34a1a73f04c03cbde9: Status 404 returned error can't find the container with id 3df37b90178a71c6b9e5d94a067f0299b4d205eaf8837d34a1a73f04c03cbde9 Dec 03 06:52:50 crc kubenswrapper[4946]: I1203 06:52:50.905254 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2q2qg" Dec 03 06:52:51 crc kubenswrapper[4946]: I1203 06:52:51.195098 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:51 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:51 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:51 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:51 crc kubenswrapper[4946]: I1203 06:52:51.195196 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:51 crc kubenswrapper[4946]: I1203 06:52:51.272488 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7b584fa1-0942-4df1-98e3-c6c4e1429c85","Type":"ContainerStarted","Data":"3df37b90178a71c6b9e5d94a067f0299b4d205eaf8837d34a1a73f04c03cbde9"} Dec 03 06:52:52 crc kubenswrapper[4946]: I1203 06:52:52.193516 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:52 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:52 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:52 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:52 crc kubenswrapper[4946]: I1203 06:52:52.193915 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:52 crc kubenswrapper[4946]: I1203 06:52:52.283794 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7b584fa1-0942-4df1-98e3-c6c4e1429c85","Type":"ContainerStarted","Data":"2c60a6b1e289fc3c6b49352709bd5405d87bae7a9ae0832d46831c5ea2b768f0"} Dec 03 06:52:53 crc kubenswrapper[4946]: I1203 06:52:53.039127 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:52:53 crc kubenswrapper[4946]: I1203 06:52:53.039217 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:52:53 crc kubenswrapper[4946]: I1203 06:52:53.192594 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:53 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:53 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:53 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:53 crc kubenswrapper[4946]: I1203 06:52:53.192662 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:53 crc kubenswrapper[4946]: I1203 06:52:53.294761 4946 generic.go:334] "Generic (PLEG): container finished" podID="7b584fa1-0942-4df1-98e3-c6c4e1429c85" containerID="2c60a6b1e289fc3c6b49352709bd5405d87bae7a9ae0832d46831c5ea2b768f0" exitCode=0 Dec 03 06:52:53 crc kubenswrapper[4946]: I1203 06:52:53.294808 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7b584fa1-0942-4df1-98e3-c6c4e1429c85","Type":"ContainerDied","Data":"2c60a6b1e289fc3c6b49352709bd5405d87bae7a9ae0832d46831c5ea2b768f0"} Dec 03 06:52:54 crc kubenswrapper[4946]: I1203 06:52:54.194706 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:54 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:54 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:54 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:54 crc kubenswrapper[4946]: I1203 06:52:54.195474 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:54 crc kubenswrapper[4946]: I1203 06:52:54.980241 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ncr4w" Dec 03 06:52:54 crc kubenswrapper[4946]: I1203 06:52:54.995478 4946 patch_prober.go:28] interesting pod/console-f9d7485db-zkld2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 03 06:52:54 crc kubenswrapper[4946]: I1203 06:52:54.995534 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zkld2" podUID="a89bae42-673b-4eb1-87d2-ef2dd919c4d4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 03 06:52:55 crc kubenswrapper[4946]: I1203 06:52:55.197058 4946 patch_prober.go:28] interesting pod/router-default-5444994796-l68r8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:52:55 crc kubenswrapper[4946]: [-]has-synced failed: reason withheld Dec 03 06:52:55 crc kubenswrapper[4946]: [+]process-running ok Dec 03 06:52:55 crc kubenswrapper[4946]: healthz check failed Dec 03 06:52:55 crc kubenswrapper[4946]: I1203 06:52:55.197206 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-l68r8" podUID="56fe8e2e-1586-4fbd-b68b-f207813a7c35" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:52:55 crc kubenswrapper[4946]: I1203 06:52:55.885447 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:55 crc kubenswrapper[4946]: I1203 06:52:55.893085 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e5abff8-4252-4371-803d-d241c81c5910-metrics-certs\") pod \"network-metrics-daemon-hkpdf\" (UID: \"2e5abff8-4252-4371-803d-d241c81c5910\") " pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:56 crc kubenswrapper[4946]: I1203 06:52:56.007061 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hkpdf" Dec 03 06:52:56 crc kubenswrapper[4946]: I1203 06:52:56.198476 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:56 crc kubenswrapper[4946]: I1203 06:52:56.204018 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-l68r8" Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.876818 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.896833 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.954099 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kubelet-dir\") pod \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\" (UID: \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\") " Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.954182 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kube-api-access\") pod \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\" (UID: \"7b584fa1-0942-4df1-98e3-c6c4e1429c85\") " Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.954239 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5add0485-955c-4aa2-8a84-6105e51f9360-kubelet-dir\") pod \"5add0485-955c-4aa2-8a84-6105e51f9360\" (UID: \"5add0485-955c-4aa2-8a84-6105e51f9360\") " Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.954289 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5add0485-955c-4aa2-8a84-6105e51f9360-kube-api-access\") pod \"5add0485-955c-4aa2-8a84-6105e51f9360\" (UID: \"5add0485-955c-4aa2-8a84-6105e51f9360\") " Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.955487 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7b584fa1-0942-4df1-98e3-c6c4e1429c85" (UID: "7b584fa1-0942-4df1-98e3-c6c4e1429c85"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.955487 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5add0485-955c-4aa2-8a84-6105e51f9360-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5add0485-955c-4aa2-8a84-6105e51f9360" (UID: "5add0485-955c-4aa2-8a84-6105e51f9360"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.959911 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7b584fa1-0942-4df1-98e3-c6c4e1429c85" (UID: "7b584fa1-0942-4df1-98e3-c6c4e1429c85"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:52:59 crc kubenswrapper[4946]: I1203 06:52:59.960077 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5add0485-955c-4aa2-8a84-6105e51f9360-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5add0485-955c-4aa2-8a84-6105e51f9360" (UID: "5add0485-955c-4aa2-8a84-6105e51f9360"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.055241 4946 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.055276 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b584fa1-0942-4df1-98e3-c6c4e1429c85-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.055287 4946 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5add0485-955c-4aa2-8a84-6105e51f9360-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.055296 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5add0485-955c-4aa2-8a84-6105e51f9360-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.355697 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7b584fa1-0942-4df1-98e3-c6c4e1429c85","Type":"ContainerDied","Data":"3df37b90178a71c6b9e5d94a067f0299b4d205eaf8837d34a1a73f04c03cbde9"} Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.355755 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3df37b90178a71c6b9e5d94a067f0299b4d205eaf8837d34a1a73f04c03cbde9" Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.355806 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.359336 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5add0485-955c-4aa2-8a84-6105e51f9360","Type":"ContainerDied","Data":"6c1dba54eb478404b44652bcacbde25e4c23953f0a2a56a3f85160ab144a3f52"} Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.359379 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c1dba54eb478404b44652bcacbde25e4c23953f0a2a56a3f85160ab144a3f52" Dec 03 06:53:00 crc kubenswrapper[4946]: I1203 06:53:00.359438 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:53:03 crc kubenswrapper[4946]: I1203 06:53:03.241118 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:53:05 crc kubenswrapper[4946]: I1203 06:53:05.002427 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:53:05 crc kubenswrapper[4946]: I1203 06:53:05.012530 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-zkld2" Dec 03 06:53:10 crc kubenswrapper[4946]: E1203 06:53:10.507148 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 06:53:10 crc kubenswrapper[4946]: E1203 06:53:10.508074 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gwqbg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hv9ql_openshift-marketplace(0ce0d8cc-befc-48b1-b330-a132440b36a4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:53:10 crc kubenswrapper[4946]: E1203 06:53:10.509420 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hv9ql" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" Dec 03 06:53:12 crc kubenswrapper[4946]: E1203 06:53:12.213374 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hv9ql" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" Dec 03 06:53:12 crc kubenswrapper[4946]: E1203 06:53:12.307012 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 06:53:12 crc kubenswrapper[4946]: E1203 06:53:12.307695 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n45sl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ch5kh_openshift-marketplace(6f544f2a-7068-4378-a9f2-3ba25efd90a9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:53:12 crc kubenswrapper[4946]: E1203 06:53:12.309050 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ch5kh" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" Dec 03 06:53:14 crc kubenswrapper[4946]: I1203 06:53:14.636886 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:53:15 crc kubenswrapper[4946]: E1203 06:53:15.203594 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ch5kh" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" Dec 03 06:53:15 crc kubenswrapper[4946]: E1203 06:53:15.280996 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 06:53:15 crc kubenswrapper[4946]: E1203 06:53:15.281138 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6gq22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6bhbz_openshift-marketplace(9e385355-ddac-4e63-935c-de1ad706600a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:53:15 crc kubenswrapper[4946]: E1203 06:53:15.282260 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-6bhbz" podUID="9e385355-ddac-4e63-935c-de1ad706600a" Dec 03 06:53:15 crc kubenswrapper[4946]: I1203 06:53:15.813585 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96d5q" Dec 03 06:53:16 crc kubenswrapper[4946]: E1203 06:53:16.287774 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6bhbz" podUID="9e385355-ddac-4e63-935c-de1ad706600a" Dec 03 06:53:16 crc kubenswrapper[4946]: E1203 06:53:16.311072 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 06:53:16 crc kubenswrapper[4946]: E1203 06:53:16.311247 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5dgzr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bd7s4_openshift-marketplace(9d975a9e-23bf-4eec-a768-0b962d4b1022): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:53:16 crc kubenswrapper[4946]: E1203 06:53:16.312369 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bd7s4" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" Dec 03 06:53:16 crc kubenswrapper[4946]: E1203 06:53:16.452242 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bd7s4" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" Dec 03 06:53:16 crc kubenswrapper[4946]: I1203 06:53:16.803777 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hkpdf"] Dec 03 06:53:16 crc kubenswrapper[4946]: W1203 06:53:16.899488 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e5abff8_4252_4371_803d_d241c81c5910.slice/crio-80ad37ff0d7a463288eeccb7bee8bccd9317d9575fe5b3cb8963739f0769d9cf WatchSource:0}: Error finding container 80ad37ff0d7a463288eeccb7bee8bccd9317d9575fe5b3cb8963739f0769d9cf: Status 404 returned error can't find the container with id 80ad37ff0d7a463288eeccb7bee8bccd9317d9575fe5b3cb8963739f0769d9cf Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.455975 4946 generic.go:334] "Generic (PLEG): container finished" podID="60435668-f809-46b8-bea8-9f2a186aa15f" containerID="04181e2c761d3f03406e98f655a602f51cd04c80f607140643191d1557a0497f" exitCode=0 Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.456154 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm4jx" event={"ID":"60435668-f809-46b8-bea8-9f2a186aa15f","Type":"ContainerDied","Data":"04181e2c761d3f03406e98f655a602f51cd04c80f607140643191d1557a0497f"} Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.460506 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" event={"ID":"2e5abff8-4252-4371-803d-d241c81c5910","Type":"ContainerStarted","Data":"2487c811e53cbf2c788a4c51e05d83a59710de83d48015b11dbe13b82e1b354b"} Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.460555 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" event={"ID":"2e5abff8-4252-4371-803d-d241c81c5910","Type":"ContainerStarted","Data":"5ae76dd668fcf2a924ed1793beb779f7981a0ec4e51df4025cba28345197cebe"} Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.460571 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hkpdf" event={"ID":"2e5abff8-4252-4371-803d-d241c81c5910","Type":"ContainerStarted","Data":"80ad37ff0d7a463288eeccb7bee8bccd9317d9575fe5b3cb8963739f0769d9cf"} Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.475206 4946 generic.go:334] "Generic (PLEG): container finished" podID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerID="0bd78a9ca41c37ca9b527254bc97ebc909577c96697a084f51c32dfe57ce454b" exitCode=0 Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.475286 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zckq8" event={"ID":"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c","Type":"ContainerDied","Data":"0bd78a9ca41c37ca9b527254bc97ebc909577c96697a084f51c32dfe57ce454b"} Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.480216 4946 generic.go:334] "Generic (PLEG): container finished" podID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerID="6ee6a72ffc756c6e7b0cca8c2fde9acf5fb2879b9785f9dfe8e4426ce1c9e829" exitCode=0 Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.480276 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bs2js" event={"ID":"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9","Type":"ContainerDied","Data":"6ee6a72ffc756c6e7b0cca8c2fde9acf5fb2879b9785f9dfe8e4426ce1c9e829"} Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.482246 4946 generic.go:334] "Generic (PLEG): container finished" podID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerID="52cdb89dfacbe0145f73ff3a42b5941f25041504c57ae6faa903a2fe678f855c" exitCode=0 Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.482269 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfzbf" event={"ID":"a48e7204-9569-494d-b0bc-af6cc81f0f4a","Type":"ContainerDied","Data":"52cdb89dfacbe0145f73ff3a42b5941f25041504c57ae6faa903a2fe678f855c"} Dec 03 06:53:17 crc kubenswrapper[4946]: I1203 06:53:17.534898 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hkpdf" podStartSLOduration=165.534882227 podStartE2EDuration="2m45.534882227s" podCreationTimestamp="2025-12-03 06:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:53:17.534317321 +0000 UTC m=+190.331007440" watchObservedRunningTime="2025-12-03 06:53:17.534882227 +0000 UTC m=+190.331572336" Dec 03 06:53:18 crc kubenswrapper[4946]: I1203 06:53:18.489773 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfzbf" event={"ID":"a48e7204-9569-494d-b0bc-af6cc81f0f4a","Type":"ContainerStarted","Data":"d8c6ce6ba794fc55cd2eaf6063a56ecb9a588e875e3a8a5657155fe53d385e88"} Dec 03 06:53:18 crc kubenswrapper[4946]: I1203 06:53:18.492990 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm4jx" event={"ID":"60435668-f809-46b8-bea8-9f2a186aa15f","Type":"ContainerStarted","Data":"e4862d50823c59989f7e58072b22623a5bdb1c4ef340b14795ef4a047a30aa1d"} Dec 03 06:53:18 crc kubenswrapper[4946]: I1203 06:53:18.494990 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zckq8" event={"ID":"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c","Type":"ContainerStarted","Data":"68345f8105d01882ff9be71304adf58ffbd32f538e824446fc53b5d41d317f9e"} Dec 03 06:53:18 crc kubenswrapper[4946]: I1203 06:53:18.497140 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bs2js" event={"ID":"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9","Type":"ContainerStarted","Data":"f75d0027f81899b5c481e428a65626311da68b319b30998bfd33e9448085aa55"} Dec 03 06:53:18 crc kubenswrapper[4946]: I1203 06:53:18.512878 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zfzbf" podStartSLOduration=1.40421901 podStartE2EDuration="35.512855872s" podCreationTimestamp="2025-12-03 06:52:43 +0000 UTC" firstStartedPulling="2025-12-03 06:52:44.121671359 +0000 UTC m=+156.918361468" lastFinishedPulling="2025-12-03 06:53:18.230308221 +0000 UTC m=+191.026998330" observedRunningTime="2025-12-03 06:53:18.5092867 +0000 UTC m=+191.305976809" watchObservedRunningTime="2025-12-03 06:53:18.512855872 +0000 UTC m=+191.309545981" Dec 03 06:53:18 crc kubenswrapper[4946]: I1203 06:53:18.529136 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fm4jx" podStartSLOduration=2.535620115 podStartE2EDuration="35.529114047s" podCreationTimestamp="2025-12-03 06:52:43 +0000 UTC" firstStartedPulling="2025-12-03 06:52:45.166012423 +0000 UTC m=+157.962702542" lastFinishedPulling="2025-12-03 06:53:18.159506365 +0000 UTC m=+190.956196474" observedRunningTime="2025-12-03 06:53:18.52676703 +0000 UTC m=+191.323457139" watchObservedRunningTime="2025-12-03 06:53:18.529114047 +0000 UTC m=+191.325804156" Dec 03 06:53:18 crc kubenswrapper[4946]: I1203 06:53:18.545514 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bs2js" podStartSLOduration=3.6909784329999997 podStartE2EDuration="32.545482306s" podCreationTimestamp="2025-12-03 06:52:46 +0000 UTC" firstStartedPulling="2025-12-03 06:52:49.252919712 +0000 UTC m=+162.049609821" lastFinishedPulling="2025-12-03 06:53:18.107423575 +0000 UTC m=+190.904113694" observedRunningTime="2025-12-03 06:53:18.542592593 +0000 UTC m=+191.339282702" watchObservedRunningTime="2025-12-03 06:53:18.545482306 +0000 UTC m=+191.342172415" Dec 03 06:53:18 crc kubenswrapper[4946]: I1203 06:53:18.564920 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zckq8" podStartSLOduration=2.594326676 podStartE2EDuration="33.564897871s" podCreationTimestamp="2025-12-03 06:52:45 +0000 UTC" firstStartedPulling="2025-12-03 06:52:47.198661798 +0000 UTC m=+159.995351907" lastFinishedPulling="2025-12-03 06:53:18.169232993 +0000 UTC m=+190.965923102" observedRunningTime="2025-12-03 06:53:18.563152921 +0000 UTC m=+191.359843030" watchObservedRunningTime="2025-12-03 06:53:18.564897871 +0000 UTC m=+191.361587980" Dec 03 06:53:23 crc kubenswrapper[4946]: I1203 06:53:23.039546 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:53:23 crc kubenswrapper[4946]: I1203 06:53:23.039881 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:53:23 crc kubenswrapper[4946]: I1203 06:53:23.276470 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-s7slv"] Dec 03 06:53:23 crc kubenswrapper[4946]: I1203 06:53:23.580851 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:53:23 crc kubenswrapper[4946]: I1203 06:53:23.581233 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:53:23 crc kubenswrapper[4946]: I1203 06:53:23.969920 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:53:23 crc kubenswrapper[4946]: I1203 06:53:23.969962 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:53:24 crc kubenswrapper[4946]: I1203 06:53:24.277402 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:53:24 crc kubenswrapper[4946]: I1203 06:53:24.277828 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:53:24 crc kubenswrapper[4946]: I1203 06:53:24.559585 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:53:24 crc kubenswrapper[4946]: I1203 06:53:24.563853 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:53:25 crc kubenswrapper[4946]: I1203 06:53:25.032812 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fm4jx"] Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.020851 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.021013 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.059477 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.535023 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fm4jx" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" containerName="registry-server" containerID="cri-o://e4862d50823c59989f7e58072b22623a5bdb1c4ef340b14795ef4a047a30aa1d" gracePeriod=2 Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.574785 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.830777 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 06:53:26 crc kubenswrapper[4946]: E1203 06:53:26.831008 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5add0485-955c-4aa2-8a84-6105e51f9360" containerName="pruner" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.831019 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="5add0485-955c-4aa2-8a84-6105e51f9360" containerName="pruner" Dec 03 06:53:26 crc kubenswrapper[4946]: E1203 06:53:26.831031 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b584fa1-0942-4df1-98e3-c6c4e1429c85" containerName="pruner" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.831036 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b584fa1-0942-4df1-98e3-c6c4e1429c85" containerName="pruner" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.831122 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b584fa1-0942-4df1-98e3-c6c4e1429c85" containerName="pruner" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.831134 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="5add0485-955c-4aa2-8a84-6105e51f9360" containerName="pruner" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.831485 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.833852 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.834091 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.843368 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.946357 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.946403 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.988910 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:53:26 crc kubenswrapper[4946]: I1203 06:53:26.989768 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.033634 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.049511 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.049573 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.049668 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.070716 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.147181 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.545825 4946 generic.go:334] "Generic (PLEG): container finished" podID="60435668-f809-46b8-bea8-9f2a186aa15f" containerID="e4862d50823c59989f7e58072b22623a5bdb1c4ef340b14795ef4a047a30aa1d" exitCode=0 Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.545908 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm4jx" event={"ID":"60435668-f809-46b8-bea8-9f2a186aa15f","Type":"ContainerDied","Data":"e4862d50823c59989f7e58072b22623a5bdb1c4ef340b14795ef4a047a30aa1d"} Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.581520 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 06:53:27 crc kubenswrapper[4946]: I1203 06:53:27.600433 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.234263 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zckq8"] Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.552862 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249","Type":"ContainerStarted","Data":"ae84279449e9b6e5501a851928f2a07556a9de041bb7fdc09bc56772b4254c0c"} Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.594704 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.694440 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-utilities\") pod \"60435668-f809-46b8-bea8-9f2a186aa15f\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.694600 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-catalog-content\") pod \"60435668-f809-46b8-bea8-9f2a186aa15f\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.694668 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jg9g\" (UniqueName: \"kubernetes.io/projected/60435668-f809-46b8-bea8-9f2a186aa15f-kube-api-access-2jg9g\") pod \"60435668-f809-46b8-bea8-9f2a186aa15f\" (UID: \"60435668-f809-46b8-bea8-9f2a186aa15f\") " Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.696583 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-utilities" (OuterVolumeSpecName: "utilities") pod "60435668-f809-46b8-bea8-9f2a186aa15f" (UID: "60435668-f809-46b8-bea8-9f2a186aa15f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.701627 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60435668-f809-46b8-bea8-9f2a186aa15f-kube-api-access-2jg9g" (OuterVolumeSpecName: "kube-api-access-2jg9g") pod "60435668-f809-46b8-bea8-9f2a186aa15f" (UID: "60435668-f809-46b8-bea8-9f2a186aa15f"). InnerVolumeSpecName "kube-api-access-2jg9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.796886 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jg9g\" (UniqueName: \"kubernetes.io/projected/60435668-f809-46b8-bea8-9f2a186aa15f-kube-api-access-2jg9g\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:28 crc kubenswrapper[4946]: I1203 06:53:28.796936 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.187231 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60435668-f809-46b8-bea8-9f2a186aa15f" (UID: "60435668-f809-46b8-bea8-9f2a186aa15f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.201928 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60435668-f809-46b8-bea8-9f2a186aa15f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.559114 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249","Type":"ContainerStarted","Data":"72914afd4e8452ae150fe8716e6736aff3bec5ce9ec9b3abadaa9b8cfb121d99"} Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.561254 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm4jx" event={"ID":"60435668-f809-46b8-bea8-9f2a186aa15f","Type":"ContainerDied","Data":"5b104fafca2ba3a19808310c17ef8a6baed66dd6fd0a662588cdbb5b6a1523b8"} Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.561300 4946 scope.go:117] "RemoveContainer" containerID="e4862d50823c59989f7e58072b22623a5bdb1c4ef340b14795ef4a047a30aa1d" Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.561366 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm4jx" Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.561420 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zckq8" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerName="registry-server" containerID="cri-o://68345f8105d01882ff9be71304adf58ffbd32f538e824446fc53b5d41d317f9e" gracePeriod=2 Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.580843 4946 scope.go:117] "RemoveContainer" containerID="04181e2c761d3f03406e98f655a602f51cd04c80f607140643191d1557a0497f" Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.590903 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fm4jx"] Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.598528 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fm4jx"] Dec 03 06:53:29 crc kubenswrapper[4946]: I1203 06:53:29.631263 4946 scope.go:117] "RemoveContainer" containerID="aca86845be52987ffcd848eb411401424d22f2e9a3fa9399459c86f66e3d51e7" Dec 03 06:53:30 crc kubenswrapper[4946]: I1203 06:53:30.570098 4946 generic.go:334] "Generic (PLEG): container finished" podID="99efd2bc-c3a2-44d4-9ca6-d4cb7a409249" containerID="72914afd4e8452ae150fe8716e6736aff3bec5ce9ec9b3abadaa9b8cfb121d99" exitCode=0 Dec 03 06:53:30 crc kubenswrapper[4946]: I1203 06:53:30.570192 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249","Type":"ContainerDied","Data":"72914afd4e8452ae150fe8716e6736aff3bec5ce9ec9b3abadaa9b8cfb121d99"} Dec 03 06:53:30 crc kubenswrapper[4946]: I1203 06:53:30.573291 4946 generic.go:334] "Generic (PLEG): container finished" podID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerID="68345f8105d01882ff9be71304adf58ffbd32f538e824446fc53b5d41d317f9e" exitCode=0 Dec 03 06:53:30 crc kubenswrapper[4946]: I1203 06:53:30.573333 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zckq8" event={"ID":"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c","Type":"ContainerDied","Data":"68345f8105d01882ff9be71304adf58ffbd32f538e824446fc53b5d41d317f9e"} Dec 03 06:53:30 crc kubenswrapper[4946]: I1203 06:53:30.631226 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bs2js"] Dec 03 06:53:30 crc kubenswrapper[4946]: I1203 06:53:30.631413 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bs2js" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerName="registry-server" containerID="cri-o://f75d0027f81899b5c481e428a65626311da68b319b30998bfd33e9448085aa55" gracePeriod=2 Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.559487 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.582525 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zckq8" event={"ID":"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c","Type":"ContainerDied","Data":"f46e52c05acb2e278cdafba8006fcf30c138e42fa92223a2e0fbf635654b9c3b"} Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.582575 4946 scope.go:117] "RemoveContainer" containerID="68345f8105d01882ff9be71304adf58ffbd32f538e824446fc53b5d41d317f9e" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.582660 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zckq8" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.593689 4946 generic.go:334] "Generic (PLEG): container finished" podID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerID="f75d0027f81899b5c481e428a65626311da68b319b30998bfd33e9448085aa55" exitCode=0 Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.600599 4946 scope.go:117] "RemoveContainer" containerID="0bd78a9ca41c37ca9b527254bc97ebc909577c96697a084f51c32dfe57ce454b" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.601172 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" path="/var/lib/kubelet/pods/60435668-f809-46b8-bea8-9f2a186aa15f/volumes" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.604126 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bs2js" event={"ID":"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9","Type":"ContainerDied","Data":"f75d0027f81899b5c481e428a65626311da68b319b30998bfd33e9448085aa55"} Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.630553 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-catalog-content\") pod \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.630853 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-utilities\") pod \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.630951 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m7lz\" (UniqueName: \"kubernetes.io/projected/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-kube-api-access-2m7lz\") pod \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\" (UID: \"7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c\") " Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.631467 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-utilities" (OuterVolumeSpecName: "utilities") pod "7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" (UID: "7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.632121 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.635290 4946 scope.go:117] "RemoveContainer" containerID="05cd91c2746f63216e4e3ad8dbed14cf39060a8687381c43551f9bf5e8e5ad55" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.636722 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-kube-api-access-2m7lz" (OuterVolumeSpecName: "kube-api-access-2m7lz") pod "7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" (UID: "7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c"). InnerVolumeSpecName "kube-api-access-2m7lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.666459 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" (UID: "7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.733422 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m7lz\" (UniqueName: \"kubernetes.io/projected/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-kube-api-access-2m7lz\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.733452 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.789884 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.914175 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zckq8"] Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.917296 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zckq8"] Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.935073 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kube-api-access\") pod \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\" (UID: \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\") " Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.935189 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kubelet-dir\") pod \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\" (UID: \"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249\") " Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.935477 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "99efd2bc-c3a2-44d4-9ca6-d4cb7a409249" (UID: "99efd2bc-c3a2-44d4-9ca6-d4cb7a409249"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:53:31 crc kubenswrapper[4946]: I1203 06:53:31.940963 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "99efd2bc-c3a2-44d4-9ca6-d4cb7a409249" (UID: "99efd2bc-c3a2-44d4-9ca6-d4cb7a409249"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.037398 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.037424 4946 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99efd2bc-c3a2-44d4-9ca6-d4cb7a409249-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.174871 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.239081 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-utilities\") pod \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.239149 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtkls\" (UniqueName: \"kubernetes.io/projected/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-kube-api-access-wtkls\") pod \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.239240 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-catalog-content\") pod \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\" (UID: \"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9\") " Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.239948 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-utilities" (OuterVolumeSpecName: "utilities") pod "63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" (UID: "63a8d2b9-3245-443a-9c0c-dcd4f56a66a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.242644 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-kube-api-access-wtkls" (OuterVolumeSpecName: "kube-api-access-wtkls") pod "63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" (UID: "63a8d2b9-3245-443a-9c0c-dcd4f56a66a9"). InnerVolumeSpecName "kube-api-access-wtkls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.338091 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" (UID: "63a8d2b9-3245-443a-9c0c-dcd4f56a66a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.340346 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtkls\" (UniqueName: \"kubernetes.io/projected/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-kube-api-access-wtkls\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.340378 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.340388 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.599835 4946 generic.go:334] "Generic (PLEG): container finished" podID="9e385355-ddac-4e63-935c-de1ad706600a" containerID="63b645acf8e13293764aa50e68cfc6de5228e0aa64690fa39b67c90929a3b351" exitCode=0 Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.599896 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bhbz" event={"ID":"9e385355-ddac-4e63-935c-de1ad706600a","Type":"ContainerDied","Data":"63b645acf8e13293764aa50e68cfc6de5228e0aa64690fa39b67c90929a3b351"} Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.601604 4946 generic.go:334] "Generic (PLEG): container finished" podID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerID="ffebb48becdeb659617b33d584f3031f42bd49d9d672c720fd2fb8bcf819ddf5" exitCode=0 Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.601665 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv9ql" event={"ID":"0ce0d8cc-befc-48b1-b330-a132440b36a4","Type":"ContainerDied","Data":"ffebb48becdeb659617b33d584f3031f42bd49d9d672c720fd2fb8bcf819ddf5"} Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.604910 4946 generic.go:334] "Generic (PLEG): container finished" podID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerID="6c2567c29138afeda4469412386ff9e5396dd071fec2e33d4d163aac2f790488" exitCode=0 Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.604972 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch5kh" event={"ID":"6f544f2a-7068-4378-a9f2-3ba25efd90a9","Type":"ContainerDied","Data":"6c2567c29138afeda4469412386ff9e5396dd071fec2e33d4d163aac2f790488"} Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.611392 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bs2js" event={"ID":"63a8d2b9-3245-443a-9c0c-dcd4f56a66a9","Type":"ContainerDied","Data":"0c5bf7c24c764c2aa1c27bba67b01febcd796299c671df13453c202a37494e91"} Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.611414 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bs2js" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.611947 4946 scope.go:117] "RemoveContainer" containerID="f75d0027f81899b5c481e428a65626311da68b319b30998bfd33e9448085aa55" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.622937 4946 generic.go:334] "Generic (PLEG): container finished" podID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerID="8ec0c5efd389154b70dd9d32944f823a247481fd3f2b5b0fe1596b8f0ae71895" exitCode=0 Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.623002 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd7s4" event={"ID":"9d975a9e-23bf-4eec-a768-0b962d4b1022","Type":"ContainerDied","Data":"8ec0c5efd389154b70dd9d32944f823a247481fd3f2b5b0fe1596b8f0ae71895"} Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.630888 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"99efd2bc-c3a2-44d4-9ca6-d4cb7a409249","Type":"ContainerDied","Data":"ae84279449e9b6e5501a851928f2a07556a9de041bb7fdc09bc56772b4254c0c"} Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.630925 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae84279449e9b6e5501a851928f2a07556a9de041bb7fdc09bc56772b4254c0c" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.630977 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.644791 4946 scope.go:117] "RemoveContainer" containerID="6ee6a72ffc756c6e7b0cca8c2fde9acf5fb2879b9785f9dfe8e4426ce1c9e829" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.674353 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bs2js"] Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.675281 4946 scope.go:117] "RemoveContainer" containerID="62abc5d6c5e73ce76fd4ff84ff1a313c1f5dc611919072bebe188667ccd0b4f0" Dec 03 06:53:32 crc kubenswrapper[4946]: I1203 06:53:32.679131 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bs2js"] Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.606641 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" path="/var/lib/kubelet/pods/63a8d2b9-3245-443a-9c0c-dcd4f56a66a9/volumes" Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.607669 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" path="/var/lib/kubelet/pods/7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c/volumes" Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.639632 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd7s4" event={"ID":"9d975a9e-23bf-4eec-a768-0b962d4b1022","Type":"ContainerStarted","Data":"fbb84895d457fa370c2cc02a3cdd9780e988df2c0ee6fc09783a29693791b759"} Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.641678 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bhbz" event={"ID":"9e385355-ddac-4e63-935c-de1ad706600a","Type":"ContainerStarted","Data":"41ed0b49165a1f3d5c1a895799715eb5b9e133623c859fc7b73361ba29f3805b"} Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.644540 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv9ql" event={"ID":"0ce0d8cc-befc-48b1-b330-a132440b36a4","Type":"ContainerStarted","Data":"fdebe15c184bc05701eb887961c2dbf563feddacde5bf699b6ad7e4daf066d4a"} Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.646585 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch5kh" event={"ID":"6f544f2a-7068-4378-a9f2-3ba25efd90a9","Type":"ContainerStarted","Data":"93cef74ffcd288c73a7d4e8fc91a75b40302ed129963858055a1c1223d544f6b"} Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.664947 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bd7s4" podStartSLOduration=2.738320667 podStartE2EDuration="48.664916582s" podCreationTimestamp="2025-12-03 06:52:45 +0000 UTC" firstStartedPulling="2025-12-03 06:52:47.195878218 +0000 UTC m=+159.992568327" lastFinishedPulling="2025-12-03 06:53:33.122474133 +0000 UTC m=+205.919164242" observedRunningTime="2025-12-03 06:53:33.663215073 +0000 UTC m=+206.459905192" watchObservedRunningTime="2025-12-03 06:53:33.664916582 +0000 UTC m=+206.461606711" Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.686990 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hv9ql" podStartSLOduration=1.757260069 podStartE2EDuration="50.686964944s" podCreationTimestamp="2025-12-03 06:52:43 +0000 UTC" firstStartedPulling="2025-12-03 06:52:44.116100399 +0000 UTC m=+156.912790508" lastFinishedPulling="2025-12-03 06:53:33.045805274 +0000 UTC m=+205.842495383" observedRunningTime="2025-12-03 06:53:33.68439207 +0000 UTC m=+206.481082189" watchObservedRunningTime="2025-12-03 06:53:33.686964944 +0000 UTC m=+206.483655063" Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.710700 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ch5kh" podStartSLOduration=2.8453633 podStartE2EDuration="50.710679244s" podCreationTimestamp="2025-12-03 06:52:43 +0000 UTC" firstStartedPulling="2025-12-03 06:52:45.139887536 +0000 UTC m=+157.936577645" lastFinishedPulling="2025-12-03 06:53:33.00520347 +0000 UTC m=+205.801893589" observedRunningTime="2025-12-03 06:53:33.702972233 +0000 UTC m=+206.499662352" watchObservedRunningTime="2025-12-03 06:53:33.710679244 +0000 UTC m=+206.507369363" Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.723283 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6bhbz" podStartSLOduration=3.922668132 podStartE2EDuration="47.723264625s" podCreationTimestamp="2025-12-03 06:52:46 +0000 UTC" firstStartedPulling="2025-12-03 06:52:49.256821164 +0000 UTC m=+162.053511273" lastFinishedPulling="2025-12-03 06:53:33.057417657 +0000 UTC m=+205.854107766" observedRunningTime="2025-12-03 06:53:33.721904816 +0000 UTC m=+206.518594935" watchObservedRunningTime="2025-12-03 06:53:33.723264625 +0000 UTC m=+206.519954744" Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.795752 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:53:33 crc kubenswrapper[4946]: I1203 06:53:33.795828 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029572 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029790 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerName="extract-content" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029801 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerName="extract-content" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029824 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerName="extract-content" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029830 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerName="extract-content" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029838 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99efd2bc-c3a2-44d4-9ca6-d4cb7a409249" containerName="pruner" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029844 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="99efd2bc-c3a2-44d4-9ca6-d4cb7a409249" containerName="pruner" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029855 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029861 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029868 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029874 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029883 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerName="extract-utilities" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029889 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerName="extract-utilities" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029902 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerName="extract-utilities" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029909 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerName="extract-utilities" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029917 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" containerName="extract-content" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029923 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" containerName="extract-content" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029931 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029937 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: E1203 06:53:34.029946 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" containerName="extract-utilities" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.029953 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" containerName="extract-utilities" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.030031 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="99efd2bc-c3a2-44d4-9ca6-d4cb7a409249" containerName="pruner" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.030044 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="63a8d2b9-3245-443a-9c0c-dcd4f56a66a9" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.030053 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ae9c92d-48b7-4970-bd37-e4e4b4b83d9c" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.030065 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="60435668-f809-46b8-bea8-9f2a186aa15f" containerName="registry-server" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.030408 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.032560 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.033291 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.047520 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.163371 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.163478 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kube-api-access\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.163560 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-var-lock\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.264268 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kube-api-access\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.264352 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-var-lock\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.264410 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.264444 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-var-lock\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.264479 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.291856 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kube-api-access\") pod \"installer-9-crc\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.346220 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.811373 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 06:53:34 crc kubenswrapper[4946]: I1203 06:53:34.846785 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-ch5kh" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="registry-server" probeResult="failure" output=< Dec 03 06:53:34 crc kubenswrapper[4946]: timeout: failed to connect service ":50051" within 1s Dec 03 06:53:34 crc kubenswrapper[4946]: > Dec 03 06:53:35 crc kubenswrapper[4946]: I1203 06:53:35.602050 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:53:35 crc kubenswrapper[4946]: I1203 06:53:35.602099 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:53:35 crc kubenswrapper[4946]: I1203 06:53:35.652677 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:53:35 crc kubenswrapper[4946]: I1203 06:53:35.658783 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b1eebbf7-f2a0-4511-9b36-fb0956b1387d","Type":"ContainerStarted","Data":"1b4a18320e8f9d553c4e4b194756941a5bd5fbb9a716fd281b010dd816f6fce0"} Dec 03 06:53:35 crc kubenswrapper[4946]: I1203 06:53:35.658837 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b1eebbf7-f2a0-4511-9b36-fb0956b1387d","Type":"ContainerStarted","Data":"76bd39c5968e1b313b5db4137486ae4a1433bee39615468a40f60d803b17a5a4"} Dec 03 06:53:35 crc kubenswrapper[4946]: I1203 06:53:35.688801 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.6887257 podStartE2EDuration="1.6887257s" podCreationTimestamp="2025-12-03 06:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:53:35.687783693 +0000 UTC m=+208.484473802" watchObservedRunningTime="2025-12-03 06:53:35.6887257 +0000 UTC m=+208.485415809" Dec 03 06:53:36 crc kubenswrapper[4946]: I1203 06:53:36.604002 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:53:36 crc kubenswrapper[4946]: I1203 06:53:36.604333 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:53:37 crc kubenswrapper[4946]: I1203 06:53:37.652263 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6bhbz" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="registry-server" probeResult="failure" output=< Dec 03 06:53:37 crc kubenswrapper[4946]: timeout: failed to connect service ":50051" within 1s Dec 03 06:53:37 crc kubenswrapper[4946]: > Dec 03 06:53:43 crc kubenswrapper[4946]: I1203 06:53:43.355628 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:53:43 crc kubenswrapper[4946]: I1203 06:53:43.356466 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:53:43 crc kubenswrapper[4946]: I1203 06:53:43.429465 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:53:43 crc kubenswrapper[4946]: I1203 06:53:43.743713 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:53:43 crc kubenswrapper[4946]: I1203 06:53:43.829425 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:53:43 crc kubenswrapper[4946]: I1203 06:53:43.867990 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:53:45 crc kubenswrapper[4946]: I1203 06:53:45.650060 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:53:46 crc kubenswrapper[4946]: I1203 06:53:46.233223 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ch5kh"] Dec 03 06:53:46 crc kubenswrapper[4946]: I1203 06:53:46.233435 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ch5kh" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="registry-server" containerID="cri-o://93cef74ffcd288c73a7d4e8fc91a75b40302ed129963858055a1c1223d544f6b" gracePeriod=2 Dec 03 06:53:46 crc kubenswrapper[4946]: I1203 06:53:46.677930 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:53:46 crc kubenswrapper[4946]: I1203 06:53:46.734823 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:53:47 crc kubenswrapper[4946]: I1203 06:53:47.743153 4946 generic.go:334] "Generic (PLEG): container finished" podID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerID="93cef74ffcd288c73a7d4e8fc91a75b40302ed129963858055a1c1223d544f6b" exitCode=0 Dec 03 06:53:47 crc kubenswrapper[4946]: I1203 06:53:47.743191 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch5kh" event={"ID":"6f544f2a-7068-4378-a9f2-3ba25efd90a9","Type":"ContainerDied","Data":"93cef74ffcd288c73a7d4e8fc91a75b40302ed129963858055a1c1223d544f6b"} Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.307880 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" podUID="f2d67105-8c6c-44b1-9baa-71424ee41402" containerName="oauth-openshift" containerID="cri-o://dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395" gracePeriod=15 Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.494673 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.591769 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-catalog-content\") pod \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.591855 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n45sl\" (UniqueName: \"kubernetes.io/projected/6f544f2a-7068-4378-a9f2-3ba25efd90a9-kube-api-access-n45sl\") pod \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.591917 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-utilities\") pod \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\" (UID: \"6f544f2a-7068-4378-a9f2-3ba25efd90a9\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.594006 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-utilities" (OuterVolumeSpecName: "utilities") pod "6f544f2a-7068-4378-a9f2-3ba25efd90a9" (UID: "6f544f2a-7068-4378-a9f2-3ba25efd90a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.600313 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f544f2a-7068-4378-a9f2-3ba25efd90a9-kube-api-access-n45sl" (OuterVolumeSpecName: "kube-api-access-n45sl") pod "6f544f2a-7068-4378-a9f2-3ba25efd90a9" (UID: "6f544f2a-7068-4378-a9f2-3ba25efd90a9"). InnerVolumeSpecName "kube-api-access-n45sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.650761 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f544f2a-7068-4378-a9f2-3ba25efd90a9" (UID: "6f544f2a-7068-4378-a9f2-3ba25efd90a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.651092 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.694001 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n45sl\" (UniqueName: \"kubernetes.io/projected/6f544f2a-7068-4378-a9f2-3ba25efd90a9-kube-api-access-n45sl\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.694051 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.694060 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f544f2a-7068-4378-a9f2-3ba25efd90a9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.749960 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ch5kh" event={"ID":"6f544f2a-7068-4378-a9f2-3ba25efd90a9","Type":"ContainerDied","Data":"9a35011ba226cb087237002e4f836f63e5a9a64f05d9f748075d2c610fb98b94"} Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.750061 4946 scope.go:117] "RemoveContainer" containerID="93cef74ffcd288c73a7d4e8fc91a75b40302ed129963858055a1c1223d544f6b" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.750094 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ch5kh" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.756609 4946 generic.go:334] "Generic (PLEG): container finished" podID="f2d67105-8c6c-44b1-9baa-71424ee41402" containerID="dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395" exitCode=0 Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.756661 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" event={"ID":"f2d67105-8c6c-44b1-9baa-71424ee41402","Type":"ContainerDied","Data":"dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395"} Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.756690 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" event={"ID":"f2d67105-8c6c-44b1-9baa-71424ee41402","Type":"ContainerDied","Data":"e5796b5e54679eae5273ab9df7a29d204f5b45c5a2ec8b20b7591a353c5594f5"} Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.756767 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-s7slv" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.782578 4946 scope.go:117] "RemoveContainer" containerID="6c2567c29138afeda4469412386ff9e5396dd071fec2e33d4d163aac2f790488" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.795970 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-provider-selection\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796054 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-login\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796107 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-cliconfig\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796125 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-policies\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796146 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-dir\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796213 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-error\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796264 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-serving-cert\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796283 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-trusted-ca-bundle\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796329 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-service-ca\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796353 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-ocp-branding-template\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796385 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-session\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796454 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-router-certs\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796509 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blgxs\" (UniqueName: \"kubernetes.io/projected/f2d67105-8c6c-44b1-9baa-71424ee41402-kube-api-access-blgxs\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.796533 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-idp-0-file-data\") pod \"f2d67105-8c6c-44b1-9baa-71424ee41402\" (UID: \"f2d67105-8c6c-44b1-9baa-71424ee41402\") " Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.797672 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.797586 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ch5kh"] Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.798155 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.798460 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.798941 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.799101 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.802602 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ch5kh"] Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.805037 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.803111 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.810563 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2d67105-8c6c-44b1-9baa-71424ee41402-kube-api-access-blgxs" (OuterVolumeSpecName: "kube-api-access-blgxs") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "kube-api-access-blgxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.811928 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.815011 4946 scope.go:117] "RemoveContainer" containerID="890a47d7b4c19a3bf12a50c521a092deca15e9345f1a940677f86b4a6cef2587" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.817792 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.817853 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.817876 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.819375 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.819387 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f2d67105-8c6c-44b1-9baa-71424ee41402" (UID: "f2d67105-8c6c-44b1-9baa-71424ee41402"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.827688 4946 scope.go:117] "RemoveContainer" containerID="dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.842138 4946 scope.go:117] "RemoveContainer" containerID="dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395" Dec 03 06:53:48 crc kubenswrapper[4946]: E1203 06:53:48.842545 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395\": container with ID starting with dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395 not found: ID does not exist" containerID="dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.842590 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395"} err="failed to get container status \"dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395\": rpc error: code = NotFound desc = could not find container \"dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395\": container with ID starting with dd9bfe7a6da876858d225fd2cc65b8807d7e8589429dad9469f2c4b5d3e75395 not found: ID does not exist" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898198 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898366 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898382 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898392 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898402 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898412 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898424 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898435 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blgxs\" (UniqueName: \"kubernetes.io/projected/f2d67105-8c6c-44b1-9baa-71424ee41402-kube-api-access-blgxs\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898447 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898458 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898472 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898485 4946 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898497 4946 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:48 crc kubenswrapper[4946]: I1203 06:53:48.898507 4946 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f2d67105-8c6c-44b1-9baa-71424ee41402-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:49 crc kubenswrapper[4946]: I1203 06:53:49.104237 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-s7slv"] Dec 03 06:53:49 crc kubenswrapper[4946]: I1203 06:53:49.107550 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-s7slv"] Dec 03 06:53:49 crc kubenswrapper[4946]: I1203 06:53:49.601189 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" path="/var/lib/kubelet/pods/6f544f2a-7068-4378-a9f2-3ba25efd90a9/volumes" Dec 03 06:53:49 crc kubenswrapper[4946]: I1203 06:53:49.602464 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2d67105-8c6c-44b1-9baa-71424ee41402" path="/var/lib/kubelet/pods/f2d67105-8c6c-44b1-9baa-71424ee41402/volumes" Dec 03 06:53:53 crc kubenswrapper[4946]: I1203 06:53:53.040839 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:53:53 crc kubenswrapper[4946]: I1203 06:53:53.041221 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:53:53 crc kubenswrapper[4946]: I1203 06:53:53.041283 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:53:53 crc kubenswrapper[4946]: I1203 06:53:53.041930 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:53:53 crc kubenswrapper[4946]: I1203 06:53:53.041996 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45" gracePeriod=600 Dec 03 06:53:54 crc kubenswrapper[4946]: I1203 06:53:54.791887 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45" exitCode=0 Dec 03 06:53:54 crc kubenswrapper[4946]: I1203 06:53:54.791969 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45"} Dec 03 06:53:54 crc kubenswrapper[4946]: I1203 06:53:54.792406 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"234eaea414d7f69a04537ca3a86c1a0eff057f57f1c02893522ad634898ab7d8"} Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.239035 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-74b487c797-7hvmz"] Dec 03 06:53:59 crc kubenswrapper[4946]: E1203 06:53:59.239693 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="registry-server" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.239705 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="registry-server" Dec 03 06:53:59 crc kubenswrapper[4946]: E1203 06:53:59.239716 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="extract-content" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.239722 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="extract-content" Dec 03 06:53:59 crc kubenswrapper[4946]: E1203 06:53:59.239750 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d67105-8c6c-44b1-9baa-71424ee41402" containerName="oauth-openshift" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.239756 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d67105-8c6c-44b1-9baa-71424ee41402" containerName="oauth-openshift" Dec 03 06:53:59 crc kubenswrapper[4946]: E1203 06:53:59.239764 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="extract-utilities" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.239770 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="extract-utilities" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.239857 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d67105-8c6c-44b1-9baa-71424ee41402" containerName="oauth-openshift" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.239867 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f544f2a-7068-4378-a9f2-3ba25efd90a9" containerName="registry-server" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.240221 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.255530 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.255552 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.257209 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.257368 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.257532 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.257525 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.257675 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.257843 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.259081 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.259192 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.259154 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.261309 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.268592 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.271981 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74b487c797-7hvmz"] Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.275529 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.276483 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333508 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-service-ca\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333548 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-router-certs\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333582 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-login\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333613 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-error\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333706 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-session\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333760 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333786 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333820 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333849 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333870 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-audit-policies\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333894 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.333980 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d3f44da-4b9c-4094-99f3-60e051c38fa2-audit-dir\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.334013 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7qjn\" (UniqueName: \"kubernetes.io/projected/0d3f44da-4b9c-4094-99f3-60e051c38fa2-kube-api-access-c7qjn\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.334037 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.434761 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-error\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.434827 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-session\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.434861 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.434884 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.434910 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.434937 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-audit-policies\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.434956 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.434979 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.435011 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d3f44da-4b9c-4094-99f3-60e051c38fa2-audit-dir\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.435036 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7qjn\" (UniqueName: \"kubernetes.io/projected/0d3f44da-4b9c-4094-99f3-60e051c38fa2-kube-api-access-c7qjn\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.435058 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.435099 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-service-ca\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.435126 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-router-certs\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.435162 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-login\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.435697 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d3f44da-4b9c-4094-99f3-60e051c38fa2-audit-dir\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.436452 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-audit-policies\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.436632 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.436715 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-service-ca\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.437022 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.440438 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-error\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.440483 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-session\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.441485 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-login\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.443265 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-router-certs\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.444362 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.445099 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.445289 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.451194 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d3f44da-4b9c-4094-99f3-60e051c38fa2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.458882 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7qjn\" (UniqueName: \"kubernetes.io/projected/0d3f44da-4b9c-4094-99f3-60e051c38fa2-kube-api-access-c7qjn\") pod \"oauth-openshift-74b487c797-7hvmz\" (UID: \"0d3f44da-4b9c-4094-99f3-60e051c38fa2\") " pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.554898 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.789611 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74b487c797-7hvmz"] Dec 03 06:53:59 crc kubenswrapper[4946]: I1203 06:53:59.826821 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" event={"ID":"0d3f44da-4b9c-4094-99f3-60e051c38fa2","Type":"ContainerStarted","Data":"07a75659b9e36478d2566d90ab8fad0e43a600b6b66321760e94d724994e3afc"} Dec 03 06:54:00 crc kubenswrapper[4946]: I1203 06:54:00.834938 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" event={"ID":"0d3f44da-4b9c-4094-99f3-60e051c38fa2","Type":"ContainerStarted","Data":"56f8e06ef9ab1bbccb2ac4b23480c8f955ca65817238feb63af8f5d91ecb1c22"} Dec 03 06:54:00 crc kubenswrapper[4946]: I1203 06:54:00.835424 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:54:00 crc kubenswrapper[4946]: I1203 06:54:00.840867 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" Dec 03 06:54:00 crc kubenswrapper[4946]: I1203 06:54:00.854341 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-74b487c797-7hvmz" podStartSLOduration=37.854323097 podStartE2EDuration="37.854323097s" podCreationTimestamp="2025-12-03 06:53:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:54:00.852441053 +0000 UTC m=+233.649131162" watchObservedRunningTime="2025-12-03 06:54:00.854323097 +0000 UTC m=+233.651013206" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.904536 4946 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.905669 4946 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.905832 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906038 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d" gracePeriod=15 Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906077 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871" gracePeriod=15 Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906153 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313" gracePeriod=15 Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906249 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41" gracePeriod=15 Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906199 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1" gracePeriod=15 Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906625 4946 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:54:12 crc kubenswrapper[4946]: E1203 06:54:12.906819 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906838 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 06:54:12 crc kubenswrapper[4946]: E1203 06:54:12.906850 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906857 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:54:12 crc kubenswrapper[4946]: E1203 06:54:12.906869 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906877 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 06:54:12 crc kubenswrapper[4946]: E1203 06:54:12.906883 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906890 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 06:54:12 crc kubenswrapper[4946]: E1203 06:54:12.906901 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906909 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 06:54:12 crc kubenswrapper[4946]: E1203 06:54:12.906921 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.906928 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.907033 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.907045 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.907055 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.907063 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.907074 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 06:54:12 crc kubenswrapper[4946]: E1203 06:54:12.907159 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.907167 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:54:12 crc kubenswrapper[4946]: I1203 06:54:12.907266 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.031198 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.031253 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.031311 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.031332 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.031382 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.031400 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.031439 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.031471 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133083 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133137 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133181 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133200 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133227 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133224 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133267 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133241 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133275 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133284 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133298 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133344 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133371 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133397 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133408 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.133449 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.913565 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.914983 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.915921 4946 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41" exitCode=0 Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.915945 4946 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871" exitCode=0 Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.915952 4946 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1" exitCode=0 Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.915959 4946 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313" exitCode=2 Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.915997 4946 scope.go:117] "RemoveContainer" containerID="7d764e4fdce996ddfbf89024dca7b9850f5087b274ad04c7038250af980f1ba4" Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.918109 4946 generic.go:334] "Generic (PLEG): container finished" podID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" containerID="1b4a18320e8f9d553c4e4b194756941a5bd5fbb9a716fd281b010dd816f6fce0" exitCode=0 Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.918139 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b1eebbf7-f2a0-4511-9b36-fb0956b1387d","Type":"ContainerDied","Data":"1b4a18320e8f9d553c4e4b194756941a5bd5fbb9a716fd281b010dd816f6fce0"} Dec 03 06:54:13 crc kubenswrapper[4946]: I1203 06:54:13.918860 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:14 crc kubenswrapper[4946]: I1203 06:54:14.931663 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.298143 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.298295 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.298854 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.299409 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.299815 4946 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.300289 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.364843 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.364918 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kubelet-dir\") pod \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.364948 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-var-lock\") pod \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365003 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.364987 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365024 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b1eebbf7-f2a0-4511-9b36-fb0956b1387d" (UID: "b1eebbf7-f2a0-4511-9b36-fb0956b1387d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365042 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-var-lock" (OuterVolumeSpecName: "var-lock") pod "b1eebbf7-f2a0-4511-9b36-fb0956b1387d" (UID: "b1eebbf7-f2a0-4511-9b36-fb0956b1387d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365058 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365065 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kube-api-access\") pod \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\" (UID: \"b1eebbf7-f2a0-4511-9b36-fb0956b1387d\") " Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365174 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365571 4946 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365590 4946 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365600 4946 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365609 4946 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.365628 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.380116 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b1eebbf7-f2a0-4511-9b36-fb0956b1387d" (UID: "b1eebbf7-f2a0-4511-9b36-fb0956b1387d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.466805 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b1eebbf7-f2a0-4511-9b36-fb0956b1387d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.466835 4946 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.604939 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 06:54:15 crc kubenswrapper[4946]: E1203 06:54:15.614030 4946 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-42855" volumeName="registry-storage" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.948337 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b1eebbf7-f2a0-4511-9b36-fb0956b1387d","Type":"ContainerDied","Data":"76bd39c5968e1b313b5db4137486ae4a1433bee39615468a40f60d803b17a5a4"} Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.948973 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76bd39c5968e1b313b5db4137486ae4a1433bee39615468a40f60d803b17a5a4" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.949278 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.954910 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.955790 4946 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d" exitCode=0 Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.955851 4946 scope.go:117] "RemoveContainer" containerID="c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.955916 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.955902 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.956285 4946 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.956524 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.959366 4946 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.959920 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.982853 4946 scope.go:117] "RemoveContainer" containerID="880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871" Dec 03 06:54:15 crc kubenswrapper[4946]: I1203 06:54:15.998005 4946 scope.go:117] "RemoveContainer" containerID="6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.016929 4946 scope.go:117] "RemoveContainer" containerID="34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.031922 4946 scope.go:117] "RemoveContainer" containerID="0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.050510 4946 scope.go:117] "RemoveContainer" containerID="fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.072205 4946 scope.go:117] "RemoveContainer" containerID="c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41" Dec 03 06:54:16 crc kubenswrapper[4946]: E1203 06:54:16.072637 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\": container with ID starting with c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41 not found: ID does not exist" containerID="c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.072683 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41"} err="failed to get container status \"c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\": rpc error: code = NotFound desc = could not find container \"c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41\": container with ID starting with c3a2967dc803fc15a259018ae07201df314aaa42f0d1608b422f894602d67c41 not found: ID does not exist" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.072711 4946 scope.go:117] "RemoveContainer" containerID="880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871" Dec 03 06:54:16 crc kubenswrapper[4946]: E1203 06:54:16.073091 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\": container with ID starting with 880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871 not found: ID does not exist" containerID="880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.073171 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871"} err="failed to get container status \"880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\": rpc error: code = NotFound desc = could not find container \"880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871\": container with ID starting with 880550716eee31bb6c0205da4d77db88834243d07aac758208e0bdbe59157871 not found: ID does not exist" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.073198 4946 scope.go:117] "RemoveContainer" containerID="6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1" Dec 03 06:54:16 crc kubenswrapper[4946]: E1203 06:54:16.073610 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\": container with ID starting with 6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1 not found: ID does not exist" containerID="6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.073631 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1"} err="failed to get container status \"6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\": rpc error: code = NotFound desc = could not find container \"6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1\": container with ID starting with 6eac20f726b3ffff882ab5ee95214ce677b74a52abb94b81e3e6de435b45b9b1 not found: ID does not exist" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.073645 4946 scope.go:117] "RemoveContainer" containerID="34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313" Dec 03 06:54:16 crc kubenswrapper[4946]: E1203 06:54:16.073991 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\": container with ID starting with 34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313 not found: ID does not exist" containerID="34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.074017 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313"} err="failed to get container status \"34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\": rpc error: code = NotFound desc = could not find container \"34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313\": container with ID starting with 34ddaa2cbb9c67b57caa734ab3869f9a546d2563f87914a3cd3a92f0586b2313 not found: ID does not exist" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.074032 4946 scope.go:117] "RemoveContainer" containerID="0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d" Dec 03 06:54:16 crc kubenswrapper[4946]: E1203 06:54:16.074304 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\": container with ID starting with 0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d not found: ID does not exist" containerID="0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.074335 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d"} err="failed to get container status \"0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\": rpc error: code = NotFound desc = could not find container \"0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d\": container with ID starting with 0c5697d5937664886dfe3c24e61b2df11ae6f7f0d1c1ec71d9cd149c9e95888d not found: ID does not exist" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.074349 4946 scope.go:117] "RemoveContainer" containerID="fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f" Dec 03 06:54:16 crc kubenswrapper[4946]: E1203 06:54:16.075051 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\": container with ID starting with fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f not found: ID does not exist" containerID="fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f" Dec 03 06:54:16 crc kubenswrapper[4946]: I1203 06:54:16.075089 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f"} err="failed to get container status \"fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\": rpc error: code = NotFound desc = could not find container \"fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f\": container with ID starting with fc488ae0bf480d0dfa4bc82c2f76cea22091ce2deddf2bd3596a4b1182bdfb0f not found: ID does not exist" Dec 03 06:54:17 crc kubenswrapper[4946]: I1203 06:54:17.594959 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:17 crc kubenswrapper[4946]: I1203 06:54:17.595503 4946 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:17 crc kubenswrapper[4946]: E1203 06:54:17.940077 4946 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:17 crc kubenswrapper[4946]: I1203 06:54:17.940481 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:17 crc kubenswrapper[4946]: W1203 06:54:17.965431 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-4a1eb8321a7945cbd38f6b8c65d33093f5166d4d731728a039be75386ee70eb3 WatchSource:0}: Error finding container 4a1eb8321a7945cbd38f6b8c65d33093f5166d4d731728a039be75386ee70eb3: Status 404 returned error can't find the container with id 4a1eb8321a7945cbd38f6b8c65d33093f5166d4d731728a039be75386ee70eb3 Dec 03 06:54:17 crc kubenswrapper[4946]: E1203 06:54:17.969799 4946 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187da2192a73721e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 06:54:17.969193502 +0000 UTC m=+250.765883611,LastTimestamp:2025-12-03 06:54:17.969193502 +0000 UTC m=+250.765883611,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 06:54:18 crc kubenswrapper[4946]: I1203 06:54:18.975751 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f"} Dec 03 06:54:18 crc kubenswrapper[4946]: I1203 06:54:18.976854 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4a1eb8321a7945cbd38f6b8c65d33093f5166d4d731728a039be75386ee70eb3"} Dec 03 06:54:18 crc kubenswrapper[4946]: E1203 06:54:18.977518 4946 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:18 crc kubenswrapper[4946]: I1203 06:54:18.977534 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:19 crc kubenswrapper[4946]: E1203 06:54:19.661097 4946 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:19 crc kubenswrapper[4946]: E1203 06:54:19.661475 4946 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:19 crc kubenswrapper[4946]: E1203 06:54:19.661843 4946 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:19 crc kubenswrapper[4946]: E1203 06:54:19.662147 4946 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:19 crc kubenswrapper[4946]: E1203 06:54:19.662396 4946 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:19 crc kubenswrapper[4946]: I1203 06:54:19.662430 4946 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 06:54:19 crc kubenswrapper[4946]: E1203 06:54:19.662719 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Dec 03 06:54:19 crc kubenswrapper[4946]: E1203 06:54:19.863500 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Dec 03 06:54:19 crc kubenswrapper[4946]: E1203 06:54:19.984240 4946 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:54:20 crc kubenswrapper[4946]: E1203 06:54:20.264237 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Dec 03 06:54:21 crc kubenswrapper[4946]: E1203 06:54:21.065218 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Dec 03 06:54:22 crc kubenswrapper[4946]: E1203 06:54:22.667122 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Dec 03 06:54:25 crc kubenswrapper[4946]: E1203 06:54:25.868659 4946 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="6.4s" Dec 03 06:54:26 crc kubenswrapper[4946]: I1203 06:54:26.592531 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:26 crc kubenswrapper[4946]: I1203 06:54:26.593518 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:26 crc kubenswrapper[4946]: I1203 06:54:26.615054 4946 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:26 crc kubenswrapper[4946]: I1203 06:54:26.615099 4946 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:26 crc kubenswrapper[4946]: E1203 06:54:26.615603 4946 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:26 crc kubenswrapper[4946]: I1203 06:54:26.616424 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:26 crc kubenswrapper[4946]: E1203 06:54:26.871548 4946 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187da2192a73721e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 06:54:17.969193502 +0000 UTC m=+250.765883611,LastTimestamp:2025-12-03 06:54:17.969193502 +0000 UTC m=+250.765883611,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.029425 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.029508 4946 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3" exitCode=1 Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.029603 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3"} Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.030168 4946 scope.go:117] "RemoveContainer" containerID="a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3" Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.031506 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.032021 4946 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.033097 4946 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="659ad3bae95b4eddf64ab36d970243f68627d9a1940b38f21ff28a63cf0f1a23" exitCode=0 Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.033175 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"659ad3bae95b4eddf64ab36d970243f68627d9a1940b38f21ff28a63cf0f1a23"} Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.033237 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ef554a030ffd55e7b8c51fb193503b5bc8ba8daaf27df6878d964bbae5a39df8"} Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.033717 4946 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.033782 4946 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:27 crc kubenswrapper[4946]: E1203 06:54:27.034203 4946 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.034270 4946 status_manager.go:851] "Failed to get status for pod" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:27 crc kubenswrapper[4946]: I1203 06:54:27.034833 4946 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 03 06:54:28 crc kubenswrapper[4946]: I1203 06:54:28.042288 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 06:54:28 crc kubenswrapper[4946]: I1203 06:54:28.042750 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d76975893dcbb54d5c96ace2449a48d594ef888ec516451b7a7014313b60aab8"} Dec 03 06:54:28 crc kubenswrapper[4946]: I1203 06:54:28.046668 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"32788969bb984c42c9b205c357b78cd6ba7d78bad749e3532458925a901a95ce"} Dec 03 06:54:28 crc kubenswrapper[4946]: I1203 06:54:28.046700 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1b197e28664bc8ac556aa52779b3fb78c549458581e39ff46b1509ae72bbd768"} Dec 03 06:54:28 crc kubenswrapper[4946]: I1203 06:54:28.046708 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"06ce06264c0df418a2d8ca17633a346abce0451eda837be03cd6b83ccbd24961"} Dec 03 06:54:29 crc kubenswrapper[4946]: I1203 06:54:29.053138 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ac5160b73071cf4f885ab9f326cf63224edb348fc6dc8f0aeed695a5e53c13ce"} Dec 03 06:54:29 crc kubenswrapper[4946]: I1203 06:54:29.053177 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fd16ff0c1e68ff105b7aaf11c229f6e62aba1ffa5a5171b0d8247115033acd78"} Dec 03 06:54:29 crc kubenswrapper[4946]: I1203 06:54:29.053341 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:29 crc kubenswrapper[4946]: I1203 06:54:29.053456 4946 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:29 crc kubenswrapper[4946]: I1203 06:54:29.053481 4946 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:30 crc kubenswrapper[4946]: I1203 06:54:30.994900 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:54:30 crc kubenswrapper[4946]: I1203 06:54:30.995040 4946 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 06:54:30 crc kubenswrapper[4946]: I1203 06:54:30.995321 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 06:54:31 crc kubenswrapper[4946]: I1203 06:54:31.616626 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:31 crc kubenswrapper[4946]: I1203 06:54:31.616706 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:31 crc kubenswrapper[4946]: I1203 06:54:31.630223 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:32 crc kubenswrapper[4946]: I1203 06:54:32.785980 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:54:34 crc kubenswrapper[4946]: I1203 06:54:34.060149 4946 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:34 crc kubenswrapper[4946]: I1203 06:54:34.085892 4946 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:34 crc kubenswrapper[4946]: I1203 06:54:34.085931 4946 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:34 crc kubenswrapper[4946]: I1203 06:54:34.094013 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:34 crc kubenswrapper[4946]: I1203 06:54:34.101140 4946 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="579222af-5e8a-46ae-839a-1e6aba342897" Dec 03 06:54:35 crc kubenswrapper[4946]: I1203 06:54:35.092671 4946 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:35 crc kubenswrapper[4946]: I1203 06:54:35.092718 4946 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="68bb29de-3574-4bd9-aa64-1da58d3dd47e" Dec 03 06:54:37 crc kubenswrapper[4946]: I1203 06:54:37.614495 4946 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="579222af-5e8a-46ae-839a-1e6aba342897" Dec 03 06:54:40 crc kubenswrapper[4946]: I1203 06:54:40.995343 4946 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 06:54:40 crc kubenswrapper[4946]: I1203 06:54:40.995843 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 06:54:43 crc kubenswrapper[4946]: I1203 06:54:43.676448 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 06:54:44 crc kubenswrapper[4946]: I1203 06:54:44.040329 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 06:54:44 crc kubenswrapper[4946]: I1203 06:54:44.192113 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 06:54:44 crc kubenswrapper[4946]: I1203 06:54:44.690831 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 06:54:45 crc kubenswrapper[4946]: I1203 06:54:45.170040 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 06:54:45 crc kubenswrapper[4946]: I1203 06:54:45.740437 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 06:54:45 crc kubenswrapper[4946]: I1203 06:54:45.812066 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 06:54:45 crc kubenswrapper[4946]: I1203 06:54:45.851420 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 06:54:45 crc kubenswrapper[4946]: I1203 06:54:45.966572 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 06:54:45 crc kubenswrapper[4946]: I1203 06:54:45.979788 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 06:54:45 crc kubenswrapper[4946]: I1203 06:54:45.995846 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.087811 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.104489 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.145648 4946 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.205521 4946 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.237229 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.259113 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.275848 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.493524 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.779592 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.799287 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.820533 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.825034 4946 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 06:54:46 crc kubenswrapper[4946]: I1203 06:54:46.940115 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 06:54:47 crc kubenswrapper[4946]: I1203 06:54:47.143326 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 06:54:47 crc kubenswrapper[4946]: I1203 06:54:47.214081 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 06:54:47 crc kubenswrapper[4946]: I1203 06:54:47.496145 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 06:54:47 crc kubenswrapper[4946]: I1203 06:54:47.498899 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 06:54:47 crc kubenswrapper[4946]: I1203 06:54:47.715687 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 06:54:47 crc kubenswrapper[4946]: I1203 06:54:47.794025 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 06:54:47 crc kubenswrapper[4946]: I1203 06:54:47.806807 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 06:54:47 crc kubenswrapper[4946]: I1203 06:54:47.812063 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.178206 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.204222 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.293974 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.337394 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.376815 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.444330 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.444353 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.464357 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.477153 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.619207 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.624003 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.656332 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.700394 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.715015 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.776749 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.796567 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.920865 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.924729 4946 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.926430 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.966662 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.969700 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.977347 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 06:54:48 crc kubenswrapper[4946]: I1203 06:54:48.999208 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.011052 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.058214 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.265841 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.311369 4946 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.316544 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.318609 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.318689 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.324723 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.325647 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.336001 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.339096 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.346783 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.346717134 podStartE2EDuration="15.346717134s" podCreationTimestamp="2025-12-03 06:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:54:49.341226628 +0000 UTC m=+282.137916787" watchObservedRunningTime="2025-12-03 06:54:49.346717134 +0000 UTC m=+282.143407293" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.354292 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.370718 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.488909 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.599425 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.661242 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.708689 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.786635 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.799429 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.958145 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.980033 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 06:54:49 crc kubenswrapper[4946]: I1203 06:54:49.992506 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.083996 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.221976 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.290961 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.322939 4946 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.348118 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.351363 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.374930 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.421083 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.497373 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.497808 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.534461 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.646565 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.657216 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.669656 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.685173 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.710157 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.735694 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.796183 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.835522 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.884344 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.993333 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.994358 4946 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.994432 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.994501 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.994567 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.995361 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"d76975893dcbb54d5c96ace2449a48d594ef888ec516451b7a7014313b60aab8"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 03 06:54:50 crc kubenswrapper[4946]: I1203 06:54:50.995559 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://d76975893dcbb54d5c96ace2449a48d594ef888ec516451b7a7014313b60aab8" gracePeriod=30 Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.047058 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.076371 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.147202 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.152564 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.175015 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.229245 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.340995 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.370828 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.429699 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.439137 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.537559 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.647833 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.669416 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.701424 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.759312 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.786238 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.877647 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.904269 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.917143 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.923435 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 06:54:51 crc kubenswrapper[4946]: I1203 06:54:51.939279 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.030022 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.167492 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.213430 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.218134 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.270044 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.361052 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.402267 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.456859 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.711272 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.736102 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.748424 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.795564 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.822619 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.981369 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 06:54:52 crc kubenswrapper[4946]: I1203 06:54:52.981377 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.143130 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.223985 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.241811 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.319998 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.333569 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.340398 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.347895 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.352392 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.365705 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.391322 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.468961 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.582675 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.600326 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.633216 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.685668 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.686664 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.788225 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.797185 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.808518 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.853243 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.854472 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.907158 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.921170 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 06:54:53 crc kubenswrapper[4946]: I1203 06:54:53.925004 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.064645 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.174138 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.187486 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.239943 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.260787 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.263920 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.266683 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.351673 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.396959 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.419858 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.492993 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.508397 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.566020 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.622083 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.660436 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 06:54:54 crc kubenswrapper[4946]: I1203 06:54:54.844569 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.009314 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.023569 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.026099 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.072948 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.157858 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.213237 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.221264 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.221409 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.267338 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.293533 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.301705 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.307282 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.335638 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.383430 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.409131 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.410444 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.447348 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.472922 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.710999 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.771664 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.796126 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.796436 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.822389 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.834213 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 06:54:55 crc kubenswrapper[4946]: I1203 06:54:55.845562 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.021220 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.117176 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.125167 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.131150 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.149940 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.174605 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.192526 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.194016 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.278260 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.307828 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.356388 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.437276 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.536933 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.663523 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.670184 4946 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.670451 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f" gracePeriod=5 Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.782449 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 06:54:56 crc kubenswrapper[4946]: I1203 06:54:56.784958 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.010287 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.079013 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.280205 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.345350 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.398852 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.530686 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.574291 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.667983 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.757060 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.801910 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 06:54:57 crc kubenswrapper[4946]: I1203 06:54:57.948252 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.114400 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.118400 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.522995 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.692225 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.700962 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.757027 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.799383 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.883276 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.925596 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 06:54:58 crc kubenswrapper[4946]: I1203 06:54:58.953781 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.041187 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.093028 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.109486 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.232911 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.344393 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.377300 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.649548 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.728854 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 06:54:59 crc kubenswrapper[4946]: I1203 06:54:59.855803 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 06:55:00 crc kubenswrapper[4946]: I1203 06:55:00.035378 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 06:55:00 crc kubenswrapper[4946]: I1203 06:55:00.237394 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 06:55:00 crc kubenswrapper[4946]: I1203 06:55:00.265704 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 06:55:00 crc kubenswrapper[4946]: I1203 06:55:00.320230 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 06:55:00 crc kubenswrapper[4946]: I1203 06:55:00.548001 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 06:55:00 crc kubenswrapper[4946]: I1203 06:55:00.637087 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 06:55:00 crc kubenswrapper[4946]: I1203 06:55:00.672861 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 06:55:00 crc kubenswrapper[4946]: I1203 06:55:00.673530 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 06:55:01 crc kubenswrapper[4946]: I1203 06:55:01.079555 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.247908 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.248242 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.280445 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.280531 4946 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f" exitCode=137 Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.280615 4946 scope.go:117] "RemoveContainer" containerID="7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.280849 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.303946 4946 scope.go:117] "RemoveContainer" containerID="7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f" Dec 03 06:55:02 crc kubenswrapper[4946]: E1203 06:55:02.304705 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f\": container with ID starting with 7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f not found: ID does not exist" containerID="7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.304782 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f"} err="failed to get container status \"7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f\": rpc error: code = NotFound desc = could not find container \"7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f\": container with ID starting with 7512cdc829e311ed1c6f67515d3648d44a3c3d9bd7f2c4bfecac87b07b87278f not found: ID does not exist" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334190 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334227 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334249 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334261 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334281 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334416 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334477 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334522 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.334587 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.335027 4946 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.335045 4946 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.335056 4946 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.335068 4946 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.345214 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:55:02 crc kubenswrapper[4946]: I1203 06:55:02.436159 4946 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:03 crc kubenswrapper[4946]: I1203 06:55:03.606064 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 06:55:07 crc kubenswrapper[4946]: I1203 06:55:07.428650 4946 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 03 06:55:13 crc kubenswrapper[4946]: I1203 06:55:13.363844 4946 generic.go:334] "Generic (PLEG): container finished" podID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerID="355012af05ffc202a7e2a63c223304b3c97e1e698e7b6a58e2666a3d8fd8ae5c" exitCode=0 Dec 03 06:55:13 crc kubenswrapper[4946]: I1203 06:55:13.363926 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" event={"ID":"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b","Type":"ContainerDied","Data":"355012af05ffc202a7e2a63c223304b3c97e1e698e7b6a58e2666a3d8fd8ae5c"} Dec 03 06:55:13 crc kubenswrapper[4946]: I1203 06:55:13.365343 4946 scope.go:117] "RemoveContainer" containerID="355012af05ffc202a7e2a63c223304b3c97e1e698e7b6a58e2666a3d8fd8ae5c" Dec 03 06:55:14 crc kubenswrapper[4946]: I1203 06:55:14.372529 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" event={"ID":"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b","Type":"ContainerStarted","Data":"4f75931ecda47c2730fa0157c83cb168bc8e719a89e2633e398038d6ccdac40e"} Dec 03 06:55:14 crc kubenswrapper[4946]: I1203 06:55:14.373420 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:55:14 crc kubenswrapper[4946]: I1203 06:55:14.377911 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:55:21 crc kubenswrapper[4946]: I1203 06:55:21.423168 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 06:55:21 crc kubenswrapper[4946]: I1203 06:55:21.426335 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 06:55:21 crc kubenswrapper[4946]: I1203 06:55:21.426422 4946 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d76975893dcbb54d5c96ace2449a48d594ef888ec516451b7a7014313b60aab8" exitCode=137 Dec 03 06:55:21 crc kubenswrapper[4946]: I1203 06:55:21.426470 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d76975893dcbb54d5c96ace2449a48d594ef888ec516451b7a7014313b60aab8"} Dec 03 06:55:21 crc kubenswrapper[4946]: I1203 06:55:21.426517 4946 scope.go:117] "RemoveContainer" containerID="a3dd009de075fa66944240d2fcf9e48e5dc821a0f5e6ea2497da3ca5a5af61b3" Dec 03 06:55:22 crc kubenswrapper[4946]: I1203 06:55:22.436372 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 06:55:22 crc kubenswrapper[4946]: I1203 06:55:22.438134 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d4796297b7acbc2300ef6676e885289a1fc76936f72569bfae96bb0605cdee5f"} Dec 03 06:55:22 crc kubenswrapper[4946]: I1203 06:55:22.785090 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:55:30 crc kubenswrapper[4946]: I1203 06:55:30.994464 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:55:30 crc kubenswrapper[4946]: I1203 06:55:30.999291 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:55:31 crc kubenswrapper[4946]: I1203 06:55:31.495024 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.346405 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-82rd5"] Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.347271 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" podUID="a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" containerName="controller-manager" containerID="cri-o://6af688b614e38e8e553cc6aed1dd1c83f1e9f9912bbe4fba19e1391109faec24" gracePeriod=30 Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.430358 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr"] Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.430615 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" podUID="f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" containerName="route-controller-manager" containerID="cri-o://2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245" gracePeriod=30 Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.548917 4946 generic.go:334] "Generic (PLEG): container finished" podID="a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" containerID="6af688b614e38e8e553cc6aed1dd1c83f1e9f9912bbe4fba19e1391109faec24" exitCode=0 Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.549001 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" event={"ID":"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5","Type":"ContainerDied","Data":"6af688b614e38e8e553cc6aed1dd1c83f1e9f9912bbe4fba19e1391109faec24"} Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.734919 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.795334 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.888873 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-serving-cert\") pod \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.888925 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjtdz\" (UniqueName: \"kubernetes.io/projected/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-kube-api-access-gjtdz\") pod \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.888974 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-proxy-ca-bundles\") pod \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.889020 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-config\") pod \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.889065 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-client-ca\") pod \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\" (UID: \"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.889573 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-client-ca" (OuterVolumeSpecName: "client-ca") pod "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" (UID: "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.889967 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" (UID: "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.890024 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-config" (OuterVolumeSpecName: "config") pod "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" (UID: "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.894888 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" (UID: "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.894913 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-kube-api-access-gjtdz" (OuterVolumeSpecName: "kube-api-access-gjtdz") pod "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" (UID: "a2a2f2fe-f951-46db-ac9f-52dfd86a60a5"). InnerVolumeSpecName "kube-api-access-gjtdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.990733 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-serving-cert\") pod \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.990875 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtlvs\" (UniqueName: \"kubernetes.io/projected/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-kube-api-access-qtlvs\") pod \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.990977 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-config\") pod \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.991061 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-client-ca\") pod \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\" (UID: \"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16\") " Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.991420 4946 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.991460 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.991479 4946 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.991496 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.991513 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjtdz\" (UniqueName: \"kubernetes.io/projected/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5-kube-api-access-gjtdz\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.991822 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-client-ca" (OuterVolumeSpecName: "client-ca") pod "f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" (UID: "f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.991910 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-config" (OuterVolumeSpecName: "config") pod "f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" (UID: "f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.995408 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-kube-api-access-qtlvs" (OuterVolumeSpecName: "kube-api-access-qtlvs") pod "f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" (UID: "f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16"). InnerVolumeSpecName "kube-api-access-qtlvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:55:41 crc kubenswrapper[4946]: I1203 06:55:41.995480 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" (UID: "f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.092322 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.092376 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtlvs\" (UniqueName: \"kubernetes.io/projected/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-kube-api-access-qtlvs\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.092421 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.092437 4946 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.558730 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" event={"ID":"a2a2f2fe-f951-46db-ac9f-52dfd86a60a5","Type":"ContainerDied","Data":"68cd3d569acc945658276c6b4a5521c40c95690c60c7e49c388af7bcf6bfd58e"} Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.558822 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-82rd5" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.558834 4946 scope.go:117] "RemoveContainer" containerID="6af688b614e38e8e553cc6aed1dd1c83f1e9f9912bbe4fba19e1391109faec24" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.560912 4946 generic.go:334] "Generic (PLEG): container finished" podID="f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" containerID="2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245" exitCode=0 Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.560960 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" event={"ID":"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16","Type":"ContainerDied","Data":"2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245"} Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.560994 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" event={"ID":"f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16","Type":"ContainerDied","Data":"b4fc0751b413e63b6a4547399c2d87d0e5396e84af3ff4264358efc6d836d04d"} Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.561380 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.593790 4946 scope.go:117] "RemoveContainer" containerID="2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.608557 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr"] Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.619164 4946 scope.go:117] "RemoveContainer" containerID="2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.619213 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kmnwr"] Dec 03 06:55:42 crc kubenswrapper[4946]: E1203 06:55:42.619836 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245\": container with ID starting with 2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245 not found: ID does not exist" containerID="2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.620080 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245"} err="failed to get container status \"2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245\": rpc error: code = NotFound desc = could not find container \"2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245\": container with ID starting with 2420d647ec857d25922f0250ea63ec2c643ac0e9b1c667d96d1a257ab7c61245 not found: ID does not exist" Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.623282 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-82rd5"] Dec 03 06:55:42 crc kubenswrapper[4946]: I1203 06:55:42.627400 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-82rd5"] Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323387 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w"] Dec 03 06:55:43 crc kubenswrapper[4946]: E1203 06:55:43.323696 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" containerName="installer" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323719 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" containerName="installer" Dec 03 06:55:43 crc kubenswrapper[4946]: E1203 06:55:43.323754 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323765 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 06:55:43 crc kubenswrapper[4946]: E1203 06:55:43.323782 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" containerName="route-controller-manager" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323791 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" containerName="route-controller-manager" Dec 03 06:55:43 crc kubenswrapper[4946]: E1203 06:55:43.323803 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" containerName="controller-manager" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323812 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" containerName="controller-manager" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323913 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" containerName="route-controller-manager" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323951 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1eebbf7-f2a0-4511-9b36-fb0956b1387d" containerName="installer" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323961 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" containerName="controller-manager" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.323976 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.324558 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.327702 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.327773 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.328081 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.328322 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.328932 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.329409 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-85c8bf77b8-82nsm"] Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.330539 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.331686 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.333225 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.334308 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.334796 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.335099 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.336095 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w"] Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.337099 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.337229 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.341839 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.344402 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85c8bf77b8-82nsm"] Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407633 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e04bc6d-c03d-4c56-b362-83f28701ecc5-client-ca\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407675 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-client-ca\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407700 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e04bc6d-c03d-4c56-b362-83f28701ecc5-serving-cert\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407729 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xlw5\" (UniqueName: \"kubernetes.io/projected/8e04bc6d-c03d-4c56-b362-83f28701ecc5-kube-api-access-8xlw5\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407760 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-config\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407780 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhktd\" (UniqueName: \"kubernetes.io/projected/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-kube-api-access-lhktd\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407796 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-serving-cert\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407815 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-proxy-ca-bundles\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.407837 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e04bc6d-c03d-4c56-b362-83f28701ecc5-config\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.508823 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-client-ca\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.508871 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e04bc6d-c03d-4c56-b362-83f28701ecc5-serving-cert\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.508905 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xlw5\" (UniqueName: \"kubernetes.io/projected/8e04bc6d-c03d-4c56-b362-83f28701ecc5-kube-api-access-8xlw5\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.508924 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-config\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.508946 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhktd\" (UniqueName: \"kubernetes.io/projected/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-kube-api-access-lhktd\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.508961 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-serving-cert\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.508982 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-proxy-ca-bundles\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.509007 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e04bc6d-c03d-4c56-b362-83f28701ecc5-config\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.509030 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e04bc6d-c03d-4c56-b362-83f28701ecc5-client-ca\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.510009 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-client-ca\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.510801 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e04bc6d-c03d-4c56-b362-83f28701ecc5-client-ca\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.512088 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-proxy-ca-bundles\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.512094 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-config\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.516069 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e04bc6d-c03d-4c56-b362-83f28701ecc5-config\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.528585 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e04bc6d-c03d-4c56-b362-83f28701ecc5-serving-cert\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.528711 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-serving-cert\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.538539 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhktd\" (UniqueName: \"kubernetes.io/projected/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-kube-api-access-lhktd\") pod \"controller-manager-85c8bf77b8-82nsm\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.538706 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xlw5\" (UniqueName: \"kubernetes.io/projected/8e04bc6d-c03d-4c56-b362-83f28701ecc5-kube-api-access-8xlw5\") pod \"route-controller-manager-6f994dfb55-s5n8w\" (UID: \"8e04bc6d-c03d-4c56-b362-83f28701ecc5\") " pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.599595 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2a2f2fe-f951-46db-ac9f-52dfd86a60a5" path="/var/lib/kubelet/pods/a2a2f2fe-f951-46db-ac9f-52dfd86a60a5/volumes" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.601505 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16" path="/var/lib/kubelet/pods/f86bb6c2-c6d0-4c6a-9801-d1d8f34a4e16/volumes" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.641921 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.654369 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:43 crc kubenswrapper[4946]: I1203 06:55:43.966300 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85c8bf77b8-82nsm"] Dec 03 06:55:43 crc kubenswrapper[4946]: W1203 06:55:43.980454 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad7dfe7c_cb73_4efb_abb2_c7629a6b50b6.slice/crio-67d34f44f5b48418bde2a1b9ab58167b63b4df23b36075469bfa666e4b04f9f8 WatchSource:0}: Error finding container 67d34f44f5b48418bde2a1b9ab58167b63b4df23b36075469bfa666e4b04f9f8: Status 404 returned error can't find the container with id 67d34f44f5b48418bde2a1b9ab58167b63b4df23b36075469bfa666e4b04f9f8 Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.015857 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w"] Dec 03 06:55:44 crc kubenswrapper[4946]: W1203 06:55:44.021623 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e04bc6d_c03d_4c56_b362_83f28701ecc5.slice/crio-e18651b3bd5a757ac318ea115f0e620183fce2b2e5e91d042dd1e071d4a180fb WatchSource:0}: Error finding container e18651b3bd5a757ac318ea115f0e620183fce2b2e5e91d042dd1e071d4a180fb: Status 404 returned error can't find the container with id e18651b3bd5a757ac318ea115f0e620183fce2b2e5e91d042dd1e071d4a180fb Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.575279 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" event={"ID":"8e04bc6d-c03d-4c56-b362-83f28701ecc5","Type":"ContainerStarted","Data":"112a42054e6a446275ec380419351872d358402c997ec2d3deb3284dcc1cb551"} Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.575607 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.575698 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" event={"ID":"8e04bc6d-c03d-4c56-b362-83f28701ecc5","Type":"ContainerStarted","Data":"e18651b3bd5a757ac318ea115f0e620183fce2b2e5e91d042dd1e071d4a180fb"} Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.577057 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" event={"ID":"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6","Type":"ContainerStarted","Data":"faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b"} Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.577114 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" event={"ID":"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6","Type":"ContainerStarted","Data":"67d34f44f5b48418bde2a1b9ab58167b63b4df23b36075469bfa666e4b04f9f8"} Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.577970 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.586385 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.602011 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" podStartSLOduration=3.601985715 podStartE2EDuration="3.601985715s" podCreationTimestamp="2025-12-03 06:55:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:55:44.594374815 +0000 UTC m=+337.391064944" watchObservedRunningTime="2025-12-03 06:55:44.601985715 +0000 UTC m=+337.398675864" Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.613106 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" podStartSLOduration=3.613086712 podStartE2EDuration="3.613086712s" podCreationTimestamp="2025-12-03 06:55:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:55:44.612838255 +0000 UTC m=+337.409528374" watchObservedRunningTime="2025-12-03 06:55:44.613086712 +0000 UTC m=+337.409776831" Dec 03 06:55:44 crc kubenswrapper[4946]: I1203 06:55:44.657221 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f994dfb55-s5n8w" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.465944 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zfzbf"] Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.466811 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zfzbf" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerName="registry-server" containerID="cri-o://d8c6ce6ba794fc55cd2eaf6063a56ecb9a588e875e3a8a5657155fe53d385e88" gracePeriod=30 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.480118 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hv9ql"] Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.480393 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hv9ql" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerName="registry-server" containerID="cri-o://fdebe15c184bc05701eb887961c2dbf563feddacde5bf699b6ad7e4daf066d4a" gracePeriod=30 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.496626 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzrcc"] Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.496884 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" containerID="cri-o://4f75931ecda47c2730fa0157c83cb168bc8e719a89e2633e398038d6ccdac40e" gracePeriod=30 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.502448 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd7s4"] Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.502643 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bd7s4" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerName="registry-server" containerID="cri-o://fbb84895d457fa370c2cc02a3cdd9780e988df2c0ee6fc09783a29693791b759" gracePeriod=30 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.508422 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6bhbz"] Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.508723 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6bhbz" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="registry-server" containerID="cri-o://41ed0b49165a1f3d5c1a895799715eb5b9e133623c859fc7b73361ba29f3805b" gracePeriod=30 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.512838 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g2gmk"] Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.513623 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.523091 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g2gmk"] Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.568996 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80417841-c6f8-4069-99bf-c98a62caa7c9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.569079 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80417841-c6f8-4069-99bf-c98a62caa7c9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.569127 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vf9z\" (UniqueName: \"kubernetes.io/projected/80417841-c6f8-4069-99bf-c98a62caa7c9-kube-api-access-5vf9z\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.670377 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80417841-c6f8-4069-99bf-c98a62caa7c9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.670462 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80417841-c6f8-4069-99bf-c98a62caa7c9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.670523 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vf9z\" (UniqueName: \"kubernetes.io/projected/80417841-c6f8-4069-99bf-c98a62caa7c9-kube-api-access-5vf9z\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.672055 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80417841-c6f8-4069-99bf-c98a62caa7c9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.680599 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80417841-c6f8-4069-99bf-c98a62caa7c9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.693547 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vf9z\" (UniqueName: \"kubernetes.io/projected/80417841-c6f8-4069-99bf-c98a62caa7c9-kube-api-access-5vf9z\") pod \"marketplace-operator-79b997595-g2gmk\" (UID: \"80417841-c6f8-4069-99bf-c98a62caa7c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.763765 4946 generic.go:334] "Generic (PLEG): container finished" podID="9e385355-ddac-4e63-935c-de1ad706600a" containerID="41ed0b49165a1f3d5c1a895799715eb5b9e133623c859fc7b73361ba29f3805b" exitCode=0 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.763939 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bhbz" event={"ID":"9e385355-ddac-4e63-935c-de1ad706600a","Type":"ContainerDied","Data":"41ed0b49165a1f3d5c1a895799715eb5b9e133623c859fc7b73361ba29f3805b"} Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.766208 4946 generic.go:334] "Generic (PLEG): container finished" podID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerID="fdebe15c184bc05701eb887961c2dbf563feddacde5bf699b6ad7e4daf066d4a" exitCode=0 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.766281 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv9ql" event={"ID":"0ce0d8cc-befc-48b1-b330-a132440b36a4","Type":"ContainerDied","Data":"fdebe15c184bc05701eb887961c2dbf563feddacde5bf699b6ad7e4daf066d4a"} Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.768186 4946 generic.go:334] "Generic (PLEG): container finished" podID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerID="d8c6ce6ba794fc55cd2eaf6063a56ecb9a588e875e3a8a5657155fe53d385e88" exitCode=0 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.768234 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfzbf" event={"ID":"a48e7204-9569-494d-b0bc-af6cc81f0f4a","Type":"ContainerDied","Data":"d8c6ce6ba794fc55cd2eaf6063a56ecb9a588e875e3a8a5657155fe53d385e88"} Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.770389 4946 generic.go:334] "Generic (PLEG): container finished" podID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerID="fbb84895d457fa370c2cc02a3cdd9780e988df2c0ee6fc09783a29693791b759" exitCode=0 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.770426 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd7s4" event={"ID":"9d975a9e-23bf-4eec-a768-0b962d4b1022","Type":"ContainerDied","Data":"fbb84895d457fa370c2cc02a3cdd9780e988df2c0ee6fc09783a29693791b759"} Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.772136 4946 generic.go:334] "Generic (PLEG): container finished" podID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerID="4f75931ecda47c2730fa0157c83cb168bc8e719a89e2633e398038d6ccdac40e" exitCode=0 Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.772160 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" event={"ID":"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b","Type":"ContainerDied","Data":"4f75931ecda47c2730fa0157c83cb168bc8e719a89e2633e398038d6ccdac40e"} Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.772186 4946 scope.go:117] "RemoveContainer" containerID="355012af05ffc202a7e2a63c223304b3c97e1e698e7b6a58e2666a3d8fd8ae5c" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.947859 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:14 crc kubenswrapper[4946]: I1203 06:56:14.952315 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.046554 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.074461 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gq22\" (UniqueName: \"kubernetes.io/projected/9e385355-ddac-4e63-935c-de1ad706600a-kube-api-access-6gq22\") pod \"9e385355-ddac-4e63-935c-de1ad706600a\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.074560 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dgzr\" (UniqueName: \"kubernetes.io/projected/9d975a9e-23bf-4eec-a768-0b962d4b1022-kube-api-access-5dgzr\") pod \"9d975a9e-23bf-4eec-a768-0b962d4b1022\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.074623 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-utilities\") pod \"9d975a9e-23bf-4eec-a768-0b962d4b1022\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.074654 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-catalog-content\") pod \"9d975a9e-23bf-4eec-a768-0b962d4b1022\" (UID: \"9d975a9e-23bf-4eec-a768-0b962d4b1022\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.074693 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-catalog-content\") pod \"9e385355-ddac-4e63-935c-de1ad706600a\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.074807 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-utilities\") pod \"9e385355-ddac-4e63-935c-de1ad706600a\" (UID: \"9e385355-ddac-4e63-935c-de1ad706600a\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.076516 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-utilities" (OuterVolumeSpecName: "utilities") pod "9e385355-ddac-4e63-935c-de1ad706600a" (UID: "9e385355-ddac-4e63-935c-de1ad706600a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.080430 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d975a9e-23bf-4eec-a768-0b962d4b1022-kube-api-access-5dgzr" (OuterVolumeSpecName: "kube-api-access-5dgzr") pod "9d975a9e-23bf-4eec-a768-0b962d4b1022" (UID: "9d975a9e-23bf-4eec-a768-0b962d4b1022"). InnerVolumeSpecName "kube-api-access-5dgzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.080480 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e385355-ddac-4e63-935c-de1ad706600a-kube-api-access-6gq22" (OuterVolumeSpecName: "kube-api-access-6gq22") pod "9e385355-ddac-4e63-935c-de1ad706600a" (UID: "9e385355-ddac-4e63-935c-de1ad706600a"). InnerVolumeSpecName "kube-api-access-6gq22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.096909 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-utilities" (OuterVolumeSpecName: "utilities") pod "9d975a9e-23bf-4eec-a768-0b962d4b1022" (UID: "9d975a9e-23bf-4eec-a768-0b962d4b1022"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.120608 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d975a9e-23bf-4eec-a768-0b962d4b1022" (UID: "9d975a9e-23bf-4eec-a768-0b962d4b1022"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.125780 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.175724 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-trusted-ca\") pod \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.176094 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-operator-metrics\") pod \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.176294 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6ls6\" (UniqueName: \"kubernetes.io/projected/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-kube-api-access-s6ls6\") pod \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\" (UID: \"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.176465 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" (UID: "fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.176815 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.176916 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gq22\" (UniqueName: \"kubernetes.io/projected/9e385355-ddac-4e63-935c-de1ad706600a-kube-api-access-6gq22\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.176995 4946 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.177082 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dgzr\" (UniqueName: \"kubernetes.io/projected/9d975a9e-23bf-4eec-a768-0b962d4b1022-kube-api-access-5dgzr\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.177206 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.177284 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d975a9e-23bf-4eec-a768-0b962d4b1022-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.179126 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" (UID: "fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.179922 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-kube-api-access-s6ls6" (OuterVolumeSpecName: "kube-api-access-s6ls6") pod "fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" (UID: "fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b"). InnerVolumeSpecName "kube-api-access-s6ls6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.213688 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e385355-ddac-4e63-935c-de1ad706600a" (UID: "9e385355-ddac-4e63-935c-de1ad706600a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.278844 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6ls6\" (UniqueName: \"kubernetes.io/projected/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-kube-api-access-s6ls6\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.278879 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e385355-ddac-4e63-935c-de1ad706600a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.278889 4946 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.391381 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g2gmk"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.446019 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.477040 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.481696 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-catalog-content\") pod \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.481770 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-utilities\") pod \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.481815 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mktw6\" (UniqueName: \"kubernetes.io/projected/a48e7204-9569-494d-b0bc-af6cc81f0f4a-kube-api-access-mktw6\") pod \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\" (UID: \"a48e7204-9569-494d-b0bc-af6cc81f0f4a\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.482477 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-utilities" (OuterVolumeSpecName: "utilities") pod "a48e7204-9569-494d-b0bc-af6cc81f0f4a" (UID: "a48e7204-9569-494d-b0bc-af6cc81f0f4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.489567 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a48e7204-9569-494d-b0bc-af6cc81f0f4a-kube-api-access-mktw6" (OuterVolumeSpecName: "kube-api-access-mktw6") pod "a48e7204-9569-494d-b0bc-af6cc81f0f4a" (UID: "a48e7204-9569-494d-b0bc-af6cc81f0f4a"). InnerVolumeSpecName "kube-api-access-mktw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.537697 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a48e7204-9569-494d-b0bc-af6cc81f0f4a" (UID: "a48e7204-9569-494d-b0bc-af6cc81f0f4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.582768 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwqbg\" (UniqueName: \"kubernetes.io/projected/0ce0d8cc-befc-48b1-b330-a132440b36a4-kube-api-access-gwqbg\") pod \"0ce0d8cc-befc-48b1-b330-a132440b36a4\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.582857 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-utilities\") pod \"0ce0d8cc-befc-48b1-b330-a132440b36a4\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.582888 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-catalog-content\") pod \"0ce0d8cc-befc-48b1-b330-a132440b36a4\" (UID: \"0ce0d8cc-befc-48b1-b330-a132440b36a4\") " Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.583056 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.583067 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48e7204-9569-494d-b0bc-af6cc81f0f4a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.583075 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mktw6\" (UniqueName: \"kubernetes.io/projected/a48e7204-9569-494d-b0bc-af6cc81f0f4a-kube-api-access-mktw6\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.584266 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-utilities" (OuterVolumeSpecName: "utilities") pod "0ce0d8cc-befc-48b1-b330-a132440b36a4" (UID: "0ce0d8cc-befc-48b1-b330-a132440b36a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.585687 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ce0d8cc-befc-48b1-b330-a132440b36a4-kube-api-access-gwqbg" (OuterVolumeSpecName: "kube-api-access-gwqbg") pod "0ce0d8cc-befc-48b1-b330-a132440b36a4" (UID: "0ce0d8cc-befc-48b1-b330-a132440b36a4"). InnerVolumeSpecName "kube-api-access-gwqbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.632286 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ce0d8cc-befc-48b1-b330-a132440b36a4" (UID: "0ce0d8cc-befc-48b1-b330-a132440b36a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.683959 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwqbg\" (UniqueName: \"kubernetes.io/projected/0ce0d8cc-befc-48b1-b330-a132440b36a4-kube-api-access-gwqbg\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.683998 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.684115 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce0d8cc-befc-48b1-b330-a132440b36a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.782279 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bhbz" event={"ID":"9e385355-ddac-4e63-935c-de1ad706600a","Type":"ContainerDied","Data":"e584718b3b0e661fad3a0fabc12582f5cc9024f3d7164cd58728af30de3e853d"} Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.782328 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bhbz" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.782335 4946 scope.go:117] "RemoveContainer" containerID="41ed0b49165a1f3d5c1a895799715eb5b9e133623c859fc7b73361ba29f3805b" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.786147 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv9ql" event={"ID":"0ce0d8cc-befc-48b1-b330-a132440b36a4","Type":"ContainerDied","Data":"14b7c7c2581e84048b2ecb3385aeb3c11c18262cc73d23b5a22a62b41a305587"} Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.786250 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hv9ql" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.788590 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" event={"ID":"80417841-c6f8-4069-99bf-c98a62caa7c9","Type":"ContainerStarted","Data":"199fe683a2bee67d7ebed5215b37f79bf944a9a0902f133e39da958fe71fff8f"} Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.788640 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" event={"ID":"80417841-c6f8-4069-99bf-c98a62caa7c9","Type":"ContainerStarted","Data":"ab69e34b176d02347b93777c090408eaf6c441eb253bfae67a1a9f4ea397018c"} Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.789091 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.791020 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zfzbf" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.791015 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfzbf" event={"ID":"a48e7204-9569-494d-b0bc-af6cc81f0f4a","Type":"ContainerDied","Data":"1d70843e6e1d2c52f5158265161d1cfd5f9818975d4239bfe4afc58e6b573587"} Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.793029 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd7s4" event={"ID":"9d975a9e-23bf-4eec-a768-0b962d4b1022","Type":"ContainerDied","Data":"9c041125aee77e945a3bf8d6c7ea38a71eb6c4cdddfbff257ddc84b46f5254af"} Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.793123 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd7s4" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.794267 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" event={"ID":"fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b","Type":"ContainerDied","Data":"975ad01a2c0d1bbb761769b518e3aa8cdf5716fbe71bafaa4505ca2711b599e6"} Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.794334 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gzrcc" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.798550 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.801806 4946 scope.go:117] "RemoveContainer" containerID="63b645acf8e13293764aa50e68cfc6de5228e0aa64690fa39b67c90929a3b351" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.810174 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-g2gmk" podStartSLOduration=1.8101584179999999 podStartE2EDuration="1.810158418s" podCreationTimestamp="2025-12-03 06:56:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:56:15.807845535 +0000 UTC m=+368.604535664" watchObservedRunningTime="2025-12-03 06:56:15.810158418 +0000 UTC m=+368.606848527" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.831135 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6bhbz"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.831265 4946 scope.go:117] "RemoveContainer" containerID="3fd32f9c91be3d53cb86a607657c9465796f34a1149dc4934cf25f29b96e82bb" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.846384 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6bhbz"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.862724 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hv9ql"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.865160 4946 scope.go:117] "RemoveContainer" containerID="fdebe15c184bc05701eb887961c2dbf563feddacde5bf699b6ad7e4daf066d4a" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.874876 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hv9ql"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.895238 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zfzbf"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.899686 4946 scope.go:117] "RemoveContainer" containerID="ffebb48becdeb659617b33d584f3031f42bd49d9d672c720fd2fb8bcf819ddf5" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.903710 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zfzbf"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.912774 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzrcc"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.916645 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzrcc"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.920216 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd7s4"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.920571 4946 scope.go:117] "RemoveContainer" containerID="e2615c32a228cb97732278adaab0a48671250280d1c635c2b3b541a1b00afdf8" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.922732 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd7s4"] Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.933938 4946 scope.go:117] "RemoveContainer" containerID="d8c6ce6ba794fc55cd2eaf6063a56ecb9a588e875e3a8a5657155fe53d385e88" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.944973 4946 scope.go:117] "RemoveContainer" containerID="52cdb89dfacbe0145f73ff3a42b5941f25041504c57ae6faa903a2fe678f855c" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.956176 4946 scope.go:117] "RemoveContainer" containerID="f131cf049caee8df568e80e621b4ed80d56def78ed9e80c326f1cc9a95cdaebe" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.966410 4946 scope.go:117] "RemoveContainer" containerID="fbb84895d457fa370c2cc02a3cdd9780e988df2c0ee6fc09783a29693791b759" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.977640 4946 scope.go:117] "RemoveContainer" containerID="8ec0c5efd389154b70dd9d32944f823a247481fd3f2b5b0fe1596b8f0ae71895" Dec 03 06:56:15 crc kubenswrapper[4946]: I1203 06:56:15.993244 4946 scope.go:117] "RemoveContainer" containerID="144f08daae3dcd0f31e0ac9e1c7bed3d2c3b9b6e7bd6de75c323aa483dc52c97" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.004893 4946 scope.go:117] "RemoveContainer" containerID="4f75931ecda47c2730fa0157c83cb168bc8e719a89e2633e398038d6ccdac40e" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682339 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m2bgp"] Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682514 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerName="extract-utilities" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682525 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerName="extract-utilities" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682536 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="extract-content" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682542 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="extract-content" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682550 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerName="extract-content" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682556 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerName="extract-content" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682562 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682567 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682576 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerName="extract-utilities" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682581 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerName="extract-utilities" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682588 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682593 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682600 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682606 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682615 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682621 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682629 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerName="extract-content" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682635 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerName="extract-content" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682642 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="extract-utilities" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682647 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="extract-utilities" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682654 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682662 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682669 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerName="extract-content" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682675 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerName="extract-content" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682683 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682688 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: E1203 06:56:16.682696 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerName="extract-utilities" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682703 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerName="extract-utilities" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682793 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682801 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682809 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682817 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682825 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" containerName="marketplace-operator" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.682836 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e385355-ddac-4e63-935c-de1ad706600a" containerName="registry-server" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.683427 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.687875 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.692671 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2bgp"] Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.798365 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90735cf4-f312-4185-adf9-eb2112863a50-catalog-content\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.798410 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lglp2\" (UniqueName: \"kubernetes.io/projected/90735cf4-f312-4185-adf9-eb2112863a50-kube-api-access-lglp2\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.798434 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90735cf4-f312-4185-adf9-eb2112863a50-utilities\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.891241 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g8kn9"] Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.893063 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.895356 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.900086 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90735cf4-f312-4185-adf9-eb2112863a50-catalog-content\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.900224 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lglp2\" (UniqueName: \"kubernetes.io/projected/90735cf4-f312-4185-adf9-eb2112863a50-kube-api-access-lglp2\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.900276 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90735cf4-f312-4185-adf9-eb2112863a50-utilities\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.902361 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90735cf4-f312-4185-adf9-eb2112863a50-catalog-content\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.903099 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90735cf4-f312-4185-adf9-eb2112863a50-utilities\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.903394 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8kn9"] Dec 03 06:56:16 crc kubenswrapper[4946]: I1203 06:56:16.945902 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lglp2\" (UniqueName: \"kubernetes.io/projected/90735cf4-f312-4185-adf9-eb2112863a50-kube-api-access-lglp2\") pod \"redhat-marketplace-m2bgp\" (UID: \"90735cf4-f312-4185-adf9-eb2112863a50\") " pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.001530 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-utilities\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.001595 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-catalog-content\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.001663 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4gsr\" (UniqueName: \"kubernetes.io/projected/1f0bfec9-3a64-4320-96f0-f262aad05298-kube-api-access-g4gsr\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.005447 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.103598 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-utilities\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.104211 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-utilities\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.104698 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-catalog-content\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.104824 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4gsr\" (UniqueName: \"kubernetes.io/projected/1f0bfec9-3a64-4320-96f0-f262aad05298-kube-api-access-g4gsr\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.105021 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-catalog-content\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.132679 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4gsr\" (UniqueName: \"kubernetes.io/projected/1f0bfec9-3a64-4320-96f0-f262aad05298-kube-api-access-g4gsr\") pod \"redhat-operators-g8kn9\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.272803 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.604101 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ce0d8cc-befc-48b1-b330-a132440b36a4" path="/var/lib/kubelet/pods/0ce0d8cc-befc-48b1-b330-a132440b36a4/volumes" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.605549 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d975a9e-23bf-4eec-a768-0b962d4b1022" path="/var/lib/kubelet/pods/9d975a9e-23bf-4eec-a768-0b962d4b1022/volumes" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.606883 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e385355-ddac-4e63-935c-de1ad706600a" path="/var/lib/kubelet/pods/9e385355-ddac-4e63-935c-de1ad706600a/volumes" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.609201 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a48e7204-9569-494d-b0bc-af6cc81f0f4a" path="/var/lib/kubelet/pods/a48e7204-9569-494d-b0bc-af6cc81f0f4a/volumes" Dec 03 06:56:17 crc kubenswrapper[4946]: I1203 06:56:17.610815 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b" path="/var/lib/kubelet/pods/fa8ec7c1-a8aa-446e-90c4-84d5dbb3ae3b/volumes" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.278018 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2bgp"] Dec 03 06:56:18 crc kubenswrapper[4946]: W1203 06:56:18.283658 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90735cf4_f312_4185_adf9_eb2112863a50.slice/crio-1f96a4a3dc584409df996b8603629948db6762d5d36c20cea89a921a7551365a WatchSource:0}: Error finding container 1f96a4a3dc584409df996b8603629948db6762d5d36c20cea89a921a7551365a: Status 404 returned error can't find the container with id 1f96a4a3dc584409df996b8603629948db6762d5d36c20cea89a921a7551365a Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.360634 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8kn9"] Dec 03 06:56:18 crc kubenswrapper[4946]: W1203 06:56:18.373218 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f0bfec9_3a64_4320_96f0_f262aad05298.slice/crio-295904b2187cd932d5b9be51d1700a2a0c0e7673ada32fb745b99dcaf81908c4 WatchSource:0}: Error finding container 295904b2187cd932d5b9be51d1700a2a0c0e7673ada32fb745b99dcaf81908c4: Status 404 returned error can't find the container with id 295904b2187cd932d5b9be51d1700a2a0c0e7673ada32fb745b99dcaf81908c4 Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.822713 4946 generic.go:334] "Generic (PLEG): container finished" podID="90735cf4-f312-4185-adf9-eb2112863a50" containerID="ac5944eb1e0f00408a6e532d12374e6f2eb87d8c46594fb8a0143cc15075ebcc" exitCode=0 Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.824508 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2bgp" event={"ID":"90735cf4-f312-4185-adf9-eb2112863a50","Type":"ContainerDied","Data":"ac5944eb1e0f00408a6e532d12374e6f2eb87d8c46594fb8a0143cc15075ebcc"} Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.824537 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2bgp" event={"ID":"90735cf4-f312-4185-adf9-eb2112863a50","Type":"ContainerStarted","Data":"1f96a4a3dc584409df996b8603629948db6762d5d36c20cea89a921a7551365a"} Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.828012 4946 generic.go:334] "Generic (PLEG): container finished" podID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerID="b69758e7a5c0c87afadbc8ccd9bb76320ebaed1353de0217c53ad2bb2ded14f6" exitCode=0 Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.828036 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8kn9" event={"ID":"1f0bfec9-3a64-4320-96f0-f262aad05298","Type":"ContainerDied","Data":"b69758e7a5c0c87afadbc8ccd9bb76320ebaed1353de0217c53ad2bb2ded14f6"} Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.828052 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8kn9" event={"ID":"1f0bfec9-3a64-4320-96f0-f262aad05298","Type":"ContainerStarted","Data":"295904b2187cd932d5b9be51d1700a2a0c0e7673ada32fb745b99dcaf81908c4"} Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.890143 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6clpd"] Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.890812 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.910052 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6clpd"] Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.925719 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-registry-tls\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.925809 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03817e55-375c-4f6a-9497-1578ae28f98c-trusted-ca\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.925937 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-bound-sa-token\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.925968 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.926006 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/03817e55-375c-4f6a-9497-1578ae28f98c-registry-certificates\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.926028 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsjb5\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-kube-api-access-dsjb5\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.926050 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/03817e55-375c-4f6a-9497-1578ae28f98c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.926107 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/03817e55-375c-4f6a-9497-1578ae28f98c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:18 crc kubenswrapper[4946]: I1203 06:56:18.967142 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.026818 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/03817e55-375c-4f6a-9497-1578ae28f98c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.026879 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-registry-tls\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.026913 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03817e55-375c-4f6a-9497-1578ae28f98c-trusted-ca\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.026975 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-bound-sa-token\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.027004 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/03817e55-375c-4f6a-9497-1578ae28f98c-registry-certificates\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.027023 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsjb5\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-kube-api-access-dsjb5\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.027050 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/03817e55-375c-4f6a-9497-1578ae28f98c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.027491 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/03817e55-375c-4f6a-9497-1578ae28f98c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.028461 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/03817e55-375c-4f6a-9497-1578ae28f98c-registry-certificates\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.029251 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03817e55-375c-4f6a-9497-1578ae28f98c-trusted-ca\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.033436 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/03817e55-375c-4f6a-9497-1578ae28f98c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.033444 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-registry-tls\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.044768 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-bound-sa-token\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.046255 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsjb5\" (UniqueName: \"kubernetes.io/projected/03817e55-375c-4f6a-9497-1578ae28f98c-kube-api-access-dsjb5\") pod \"image-registry-66df7c8f76-6clpd\" (UID: \"03817e55-375c-4f6a-9497-1578ae28f98c\") " pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.082954 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-29nz7"] Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.094547 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-29nz7"] Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.094658 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.117659 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.135356 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-catalog-content\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.135474 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c59tl\" (UniqueName: \"kubernetes.io/projected/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-kube-api-access-c59tl\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.135533 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-utilities\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.207494 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.236915 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-catalog-content\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.237227 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c59tl\" (UniqueName: \"kubernetes.io/projected/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-kube-api-access-c59tl\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.237265 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-utilities\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.237388 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-catalog-content\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.237536 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-utilities\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.254755 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c59tl\" (UniqueName: \"kubernetes.io/projected/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-kube-api-access-c59tl\") pod \"certified-operators-29nz7\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.283162 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tww9n"] Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.284175 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.291501 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.295955 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tww9n"] Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.430842 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.440318 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-catalog-content\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.440372 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-utilities\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.440796 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb5xz\" (UniqueName: \"kubernetes.io/projected/95456701-a430-44aa-9f6a-cf6daeb3e0e0-kube-api-access-xb5xz\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.542152 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb5xz\" (UniqueName: \"kubernetes.io/projected/95456701-a430-44aa-9f6a-cf6daeb3e0e0-kube-api-access-xb5xz\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.542243 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-catalog-content\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.542269 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-utilities\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.542760 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-utilities\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.542860 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-catalog-content\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.561661 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb5xz\" (UniqueName: \"kubernetes.io/projected/95456701-a430-44aa-9f6a-cf6daeb3e0e0-kube-api-access-xb5xz\") pod \"community-operators-tww9n\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.603030 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.624655 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6clpd"] Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.847358 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8kn9" event={"ID":"1f0bfec9-3a64-4320-96f0-f262aad05298","Type":"ContainerStarted","Data":"a445de542ad83bc6bf7796fdc0de31f590cfeb65f5fecef23c010a04b7fa5190"} Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.849581 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" event={"ID":"03817e55-375c-4f6a-9497-1578ae28f98c","Type":"ContainerStarted","Data":"0db1db6ef42b3d5f70ac2c23459df31033122730fcd2274c62e764b774f30fa7"} Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.852687 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2bgp" event={"ID":"90735cf4-f312-4185-adf9-eb2112863a50","Type":"ContainerStarted","Data":"4023e69620a9fe56015cba569d9d6a33eea9f172619fed1d4b27b4b4e911391b"} Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.899932 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tww9n"] Dec 03 06:56:19 crc kubenswrapper[4946]: W1203 06:56:19.909833 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95456701_a430_44aa_9f6a_cf6daeb3e0e0.slice/crio-81303a5926d8ccf7dc25a61404a6c5e45e4a7505ed42957df55557a6f0590bfe WatchSource:0}: Error finding container 81303a5926d8ccf7dc25a61404a6c5e45e4a7505ed42957df55557a6f0590bfe: Status 404 returned error can't find the container with id 81303a5926d8ccf7dc25a61404a6c5e45e4a7505ed42957df55557a6f0590bfe Dec 03 06:56:19 crc kubenswrapper[4946]: I1203 06:56:19.978036 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-29nz7"] Dec 03 06:56:20 crc kubenswrapper[4946]: W1203 06:56:20.002663 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1f9a89f_9eef_475c_8d1c_3fcf430187cf.slice/crio-906e3f55961b7dd2b0b277fc40efba13dd7b7dce2081c41e50a184031db6b10a WatchSource:0}: Error finding container 906e3f55961b7dd2b0b277fc40efba13dd7b7dce2081c41e50a184031db6b10a: Status 404 returned error can't find the container with id 906e3f55961b7dd2b0b277fc40efba13dd7b7dce2081c41e50a184031db6b10a Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.247243 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-85c8bf77b8-82nsm"] Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.247481 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" podUID="ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" containerName="controller-manager" containerID="cri-o://faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b" gracePeriod=30 Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.571245 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.661542 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-config\") pod \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.661691 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-proxy-ca-bundles\") pod \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.661730 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhktd\" (UniqueName: \"kubernetes.io/projected/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-kube-api-access-lhktd\") pod \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.661780 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-serving-cert\") pod \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.661804 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-client-ca\") pod \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\" (UID: \"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6\") " Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.662716 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" (UID: "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.662767 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-client-ca" (OuterVolumeSpecName: "client-ca") pod "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" (UID: "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.662914 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-config" (OuterVolumeSpecName: "config") pod "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" (UID: "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.668154 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-kube-api-access-lhktd" (OuterVolumeSpecName: "kube-api-access-lhktd") pod "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" (UID: "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6"). InnerVolumeSpecName "kube-api-access-lhktd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.668776 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" (UID: "ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.764179 4946 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.764236 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhktd\" (UniqueName: \"kubernetes.io/projected/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-kube-api-access-lhktd\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.764320 4946 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.764923 4946 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.764942 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.863569 4946 generic.go:334] "Generic (PLEG): container finished" podID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerID="7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187" exitCode=0 Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.863639 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nz7" event={"ID":"b1f9a89f-9eef-475c-8d1c-3fcf430187cf","Type":"ContainerDied","Data":"7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.863664 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nz7" event={"ID":"b1f9a89f-9eef-475c-8d1c-3fcf430187cf","Type":"ContainerStarted","Data":"906e3f55961b7dd2b0b277fc40efba13dd7b7dce2081c41e50a184031db6b10a"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.866596 4946 generic.go:334] "Generic (PLEG): container finished" podID="90735cf4-f312-4185-adf9-eb2112863a50" containerID="4023e69620a9fe56015cba569d9d6a33eea9f172619fed1d4b27b4b4e911391b" exitCode=0 Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.866675 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2bgp" event={"ID":"90735cf4-f312-4185-adf9-eb2112863a50","Type":"ContainerDied","Data":"4023e69620a9fe56015cba569d9d6a33eea9f172619fed1d4b27b4b4e911391b"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.868304 4946 generic.go:334] "Generic (PLEG): container finished" podID="ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" containerID="faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b" exitCode=0 Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.868350 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.868374 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" event={"ID":"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6","Type":"ContainerDied","Data":"faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.868400 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85c8bf77b8-82nsm" event={"ID":"ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6","Type":"ContainerDied","Data":"67d34f44f5b48418bde2a1b9ab58167b63b4df23b36075469bfa666e4b04f9f8"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.868430 4946 scope.go:117] "RemoveContainer" containerID="faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.875405 4946 generic.go:334] "Generic (PLEG): container finished" podID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerID="a445de542ad83bc6bf7796fdc0de31f590cfeb65f5fecef23c010a04b7fa5190" exitCode=0 Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.875478 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8kn9" event={"ID":"1f0bfec9-3a64-4320-96f0-f262aad05298","Type":"ContainerDied","Data":"a445de542ad83bc6bf7796fdc0de31f590cfeb65f5fecef23c010a04b7fa5190"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.882815 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" event={"ID":"03817e55-375c-4f6a-9497-1578ae28f98c","Type":"ContainerStarted","Data":"ac77ff7acd9da2dd95af85eb9a40e24925789267278d48572af653683d7c633e"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.882932 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.888216 4946 generic.go:334] "Generic (PLEG): container finished" podID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerID="5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f" exitCode=0 Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.888250 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tww9n" event={"ID":"95456701-a430-44aa-9f6a-cf6daeb3e0e0","Type":"ContainerDied","Data":"5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.888290 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tww9n" event={"ID":"95456701-a430-44aa-9f6a-cf6daeb3e0e0","Type":"ContainerStarted","Data":"81303a5926d8ccf7dc25a61404a6c5e45e4a7505ed42957df55557a6f0590bfe"} Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.897862 4946 scope.go:117] "RemoveContainer" containerID="faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b" Dec 03 06:56:20 crc kubenswrapper[4946]: E1203 06:56:20.898250 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b\": container with ID starting with faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b not found: ID does not exist" containerID="faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.898278 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b"} err="failed to get container status \"faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b\": rpc error: code = NotFound desc = could not find container \"faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b\": container with ID starting with faca7c2280a86d512efb770ded79fe69619263e8e1bb72d73fb5b43a1544ec6b not found: ID does not exist" Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.917520 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-85c8bf77b8-82nsm"] Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.921587 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-85c8bf77b8-82nsm"] Dec 03 06:56:20 crc kubenswrapper[4946]: I1203 06:56:20.932493 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" podStartSLOduration=2.932477865 podStartE2EDuration="2.932477865s" podCreationTimestamp="2025-12-03 06:56:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:56:20.930289464 +0000 UTC m=+373.726979573" watchObservedRunningTime="2025-12-03 06:56:20.932477865 +0000 UTC m=+373.729167974" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.345577 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq"] Dec 03 06:56:21 crc kubenswrapper[4946]: E1203 06:56:21.346118 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" containerName="controller-manager" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.346131 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" containerName="controller-manager" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.346242 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" containerName="controller-manager" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.346602 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.356728 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.358280 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.358484 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.359581 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.360236 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.360343 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.361589 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.370575 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq"] Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.476131 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-proxy-ca-bundles\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.476171 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-config\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.476211 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwz2j\" (UniqueName: \"kubernetes.io/projected/4bac513d-dea3-4946-8760-003f4a37a73a-kube-api-access-zwz2j\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.476244 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bac513d-dea3-4946-8760-003f4a37a73a-serving-cert\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.476275 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-client-ca\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.577554 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-proxy-ca-bundles\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.577609 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-config\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.577665 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwz2j\" (UniqueName: \"kubernetes.io/projected/4bac513d-dea3-4946-8760-003f4a37a73a-kube-api-access-zwz2j\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.577725 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bac513d-dea3-4946-8760-003f4a37a73a-serving-cert\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.577800 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-client-ca\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.578864 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-client-ca\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.579503 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-proxy-ca-bundles\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.580214 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bac513d-dea3-4946-8760-003f4a37a73a-config\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.584779 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bac513d-dea3-4946-8760-003f4a37a73a-serving-cert\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.599719 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwz2j\" (UniqueName: \"kubernetes.io/projected/4bac513d-dea3-4946-8760-003f4a37a73a-kube-api-access-zwz2j\") pod \"controller-manager-6d4c657bc8-cfxhq\" (UID: \"4bac513d-dea3-4946-8760-003f4a37a73a\") " pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.603576 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6" path="/var/lib/kubelet/pods/ad7dfe7c-cb73-4efb-abb2-c7629a6b50b6/volumes" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.703329 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.896022 4946 generic.go:334] "Generic (PLEG): container finished" podID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerID="422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6" exitCode=0 Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.896325 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nz7" event={"ID":"b1f9a89f-9eef-475c-8d1c-3fcf430187cf","Type":"ContainerDied","Data":"422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6"} Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.904477 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2bgp" event={"ID":"90735cf4-f312-4185-adf9-eb2112863a50","Type":"ContainerStarted","Data":"5bf23e64222896c1127ba49bba93700a8d74c04ec4d0290d11926cb457946a40"} Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.915174 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8kn9" event={"ID":"1f0bfec9-3a64-4320-96f0-f262aad05298","Type":"ContainerStarted","Data":"6fd911556a7b0edb06aa8bc89a38946cf1361756b0a011df9881a47247e98c7c"} Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.931555 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m2bgp" podStartSLOduration=4.008950661 podStartE2EDuration="5.931509282s" podCreationTimestamp="2025-12-03 06:56:16 +0000 UTC" firstStartedPulling="2025-12-03 06:56:18.825621987 +0000 UTC m=+371.622312096" lastFinishedPulling="2025-12-03 06:56:20.748180608 +0000 UTC m=+373.544870717" observedRunningTime="2025-12-03 06:56:21.927270445 +0000 UTC m=+374.723960574" watchObservedRunningTime="2025-12-03 06:56:21.931509282 +0000 UTC m=+374.728199391" Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.969617 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g8kn9" podStartSLOduration=3.496799475 podStartE2EDuration="5.969597914s" podCreationTimestamp="2025-12-03 06:56:16 +0000 UTC" firstStartedPulling="2025-12-03 06:56:18.829177065 +0000 UTC m=+371.625867174" lastFinishedPulling="2025-12-03 06:56:21.301975504 +0000 UTC m=+374.098665613" observedRunningTime="2025-12-03 06:56:21.951256597 +0000 UTC m=+374.747946716" watchObservedRunningTime="2025-12-03 06:56:21.969597914 +0000 UTC m=+374.766288023" Dec 03 06:56:21 crc kubenswrapper[4946]: W1203 06:56:21.969769 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bac513d_dea3_4946_8760_003f4a37a73a.slice/crio-ca5f2a3b1042982fa2a2cf1f44555110774eb9ade7757dd01e7b3d135fa5a720 WatchSource:0}: Error finding container ca5f2a3b1042982fa2a2cf1f44555110774eb9ade7757dd01e7b3d135fa5a720: Status 404 returned error can't find the container with id ca5f2a3b1042982fa2a2cf1f44555110774eb9ade7757dd01e7b3d135fa5a720 Dec 03 06:56:21 crc kubenswrapper[4946]: I1203 06:56:21.991221 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq"] Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.923069 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nz7" event={"ID":"b1f9a89f-9eef-475c-8d1c-3fcf430187cf","Type":"ContainerStarted","Data":"040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5"} Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.924940 4946 generic.go:334] "Generic (PLEG): container finished" podID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerID="2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8" exitCode=0 Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.925021 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tww9n" event={"ID":"95456701-a430-44aa-9f6a-cf6daeb3e0e0","Type":"ContainerDied","Data":"2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8"} Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.926612 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" event={"ID":"4bac513d-dea3-4946-8760-003f4a37a73a","Type":"ContainerStarted","Data":"4f50cfbe19d1870f5ae797610ad1baaef6ced0ff39447d43295ae459fb504c48"} Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.926661 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" event={"ID":"4bac513d-dea3-4946-8760-003f4a37a73a","Type":"ContainerStarted","Data":"ca5f2a3b1042982fa2a2cf1f44555110774eb9ade7757dd01e7b3d135fa5a720"} Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.927014 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.930869 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.945819 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-29nz7" podStartSLOduration=2.143085569 podStartE2EDuration="3.945798341s" podCreationTimestamp="2025-12-03 06:56:19 +0000 UTC" firstStartedPulling="2025-12-03 06:56:20.866239027 +0000 UTC m=+373.662929136" lastFinishedPulling="2025-12-03 06:56:22.668951799 +0000 UTC m=+375.465641908" observedRunningTime="2025-12-03 06:56:22.944689941 +0000 UTC m=+375.741380050" watchObservedRunningTime="2025-12-03 06:56:22.945798341 +0000 UTC m=+375.742488450" Dec 03 06:56:22 crc kubenswrapper[4946]: I1203 06:56:22.989699 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d4c657bc8-cfxhq" podStartSLOduration=2.989678782 podStartE2EDuration="2.989678782s" podCreationTimestamp="2025-12-03 06:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:56:22.987544194 +0000 UTC m=+375.784234303" watchObservedRunningTime="2025-12-03 06:56:22.989678782 +0000 UTC m=+375.786368891" Dec 03 06:56:23 crc kubenswrapper[4946]: I1203 06:56:23.039662 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:56:23 crc kubenswrapper[4946]: I1203 06:56:23.039721 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:56:23 crc kubenswrapper[4946]: I1203 06:56:23.934328 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tww9n" event={"ID":"95456701-a430-44aa-9f6a-cf6daeb3e0e0","Type":"ContainerStarted","Data":"123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752"} Dec 03 06:56:27 crc kubenswrapper[4946]: I1203 06:56:27.006126 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:27 crc kubenswrapper[4946]: I1203 06:56:27.006182 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:27 crc kubenswrapper[4946]: I1203 06:56:27.056292 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:27 crc kubenswrapper[4946]: I1203 06:56:27.078651 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tww9n" podStartSLOduration=5.511257973 podStartE2EDuration="8.078619324s" podCreationTimestamp="2025-12-03 06:56:19 +0000 UTC" firstStartedPulling="2025-12-03 06:56:20.897972012 +0000 UTC m=+373.694662121" lastFinishedPulling="2025-12-03 06:56:23.465333363 +0000 UTC m=+376.262023472" observedRunningTime="2025-12-03 06:56:23.95681232 +0000 UTC m=+376.753502429" watchObservedRunningTime="2025-12-03 06:56:27.078619324 +0000 UTC m=+379.875309433" Dec 03 06:56:27 crc kubenswrapper[4946]: I1203 06:56:27.273610 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:27 crc kubenswrapper[4946]: I1203 06:56:27.273776 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:27 crc kubenswrapper[4946]: I1203 06:56:27.318360 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:28 crc kubenswrapper[4946]: I1203 06:56:28.008196 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m2bgp" Dec 03 06:56:28 crc kubenswrapper[4946]: I1203 06:56:28.014240 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 06:56:29 crc kubenswrapper[4946]: I1203 06:56:29.431660 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:29 crc kubenswrapper[4946]: I1203 06:56:29.431720 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:29 crc kubenswrapper[4946]: I1203 06:56:29.465998 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:29 crc kubenswrapper[4946]: I1203 06:56:29.604466 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:29 crc kubenswrapper[4946]: I1203 06:56:29.604520 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:29 crc kubenswrapper[4946]: I1203 06:56:29.674284 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:30 crc kubenswrapper[4946]: I1203 06:56:30.037179 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-29nz7" Dec 03 06:56:30 crc kubenswrapper[4946]: I1203 06:56:30.045508 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tww9n" Dec 03 06:56:39 crc kubenswrapper[4946]: I1203 06:56:39.212219 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-6clpd" Dec 03 06:56:39 crc kubenswrapper[4946]: I1203 06:56:39.278236 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-42855"] Dec 03 06:56:53 crc kubenswrapper[4946]: I1203 06:56:53.040295 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:56:53 crc kubenswrapper[4946]: I1203 06:56:53.041113 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.323144 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-42855" podUID="8b8c7afc-56fb-49ab-ae82-e60167809f15" containerName="registry" containerID="cri-o://e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a" gracePeriod=30 Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.743188 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.926521 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b8c7afc-56fb-49ab-ae82-e60167809f15-installation-pull-secrets\") pod \"8b8c7afc-56fb-49ab-ae82-e60167809f15\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.926900 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b8c7afc-56fb-49ab-ae82-e60167809f15-ca-trust-extracted\") pod \"8b8c7afc-56fb-49ab-ae82-e60167809f15\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.927351 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8b8c7afc-56fb-49ab-ae82-e60167809f15\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.927378 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-bound-sa-token\") pod \"8b8c7afc-56fb-49ab-ae82-e60167809f15\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.927444 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-tls\") pod \"8b8c7afc-56fb-49ab-ae82-e60167809f15\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.927498 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgxqp\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-kube-api-access-xgxqp\") pod \"8b8c7afc-56fb-49ab-ae82-e60167809f15\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.927521 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-trusted-ca\") pod \"8b8c7afc-56fb-49ab-ae82-e60167809f15\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.927554 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-certificates\") pod \"8b8c7afc-56fb-49ab-ae82-e60167809f15\" (UID: \"8b8c7afc-56fb-49ab-ae82-e60167809f15\") " Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.928496 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8b8c7afc-56fb-49ab-ae82-e60167809f15" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.928635 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8b8c7afc-56fb-49ab-ae82-e60167809f15" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.939293 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-kube-api-access-xgxqp" (OuterVolumeSpecName: "kube-api-access-xgxqp") pod "8b8c7afc-56fb-49ab-ae82-e60167809f15" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15"). InnerVolumeSpecName "kube-api-access-xgxqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.939826 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8b8c7afc-56fb-49ab-ae82-e60167809f15" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.940484 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8b8c7afc-56fb-49ab-ae82-e60167809f15" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.940490 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8b8c7afc-56fb-49ab-ae82-e60167809f15" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.942620 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b8c7afc-56fb-49ab-ae82-e60167809f15-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8b8c7afc-56fb-49ab-ae82-e60167809f15" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:57:04 crc kubenswrapper[4946]: I1203 06:57:04.952496 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b8c7afc-56fb-49ab-ae82-e60167809f15-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8b8c7afc-56fb-49ab-ae82-e60167809f15" (UID: "8b8c7afc-56fb-49ab-ae82-e60167809f15"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.029386 4946 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.029431 4946 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.029449 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgxqp\" (UniqueName: \"kubernetes.io/projected/8b8c7afc-56fb-49ab-ae82-e60167809f15-kube-api-access-xgxqp\") on node \"crc\" DevicePath \"\"" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.029466 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.029481 4946 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b8c7afc-56fb-49ab-ae82-e60167809f15-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.029500 4946 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b8c7afc-56fb-49ab-ae82-e60167809f15-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.029514 4946 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b8c7afc-56fb-49ab-ae82-e60167809f15-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.185826 4946 generic.go:334] "Generic (PLEG): container finished" podID="8b8c7afc-56fb-49ab-ae82-e60167809f15" containerID="e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a" exitCode=0 Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.185876 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-42855" event={"ID":"8b8c7afc-56fb-49ab-ae82-e60167809f15","Type":"ContainerDied","Data":"e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a"} Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.185911 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-42855" event={"ID":"8b8c7afc-56fb-49ab-ae82-e60167809f15","Type":"ContainerDied","Data":"cd212aeffb2907e8043514a3141d5e7e8ec9e03ce3cba83d8ef1a446dcf78cd3"} Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.185952 4946 scope.go:117] "RemoveContainer" containerID="e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.186220 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-42855" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.226338 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-42855"] Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.231181 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-42855"] Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.233169 4946 scope.go:117] "RemoveContainer" containerID="e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a" Dec 03 06:57:05 crc kubenswrapper[4946]: E1203 06:57:05.233601 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a\": container with ID starting with e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a not found: ID does not exist" containerID="e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.233630 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a"} err="failed to get container status \"e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a\": rpc error: code = NotFound desc = could not find container \"e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a\": container with ID starting with e18a701f7ed8cb38558fbcaa96308115c2515192c231744f4585b7f5a945b19a not found: ID does not exist" Dec 03 06:57:05 crc kubenswrapper[4946]: I1203 06:57:05.599972 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b8c7afc-56fb-49ab-ae82-e60167809f15" path="/var/lib/kubelet/pods/8b8c7afc-56fb-49ab-ae82-e60167809f15/volumes" Dec 03 06:57:23 crc kubenswrapper[4946]: I1203 06:57:23.039512 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:57:23 crc kubenswrapper[4946]: I1203 06:57:23.040241 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:57:23 crc kubenswrapper[4946]: I1203 06:57:23.041009 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 06:57:23 crc kubenswrapper[4946]: I1203 06:57:23.042060 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"234eaea414d7f69a04537ca3a86c1a0eff057f57f1c02893522ad634898ab7d8"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:57:23 crc kubenswrapper[4946]: I1203 06:57:23.042308 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://234eaea414d7f69a04537ca3a86c1a0eff057f57f1c02893522ad634898ab7d8" gracePeriod=600 Dec 03 06:57:23 crc kubenswrapper[4946]: I1203 06:57:23.297912 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="234eaea414d7f69a04537ca3a86c1a0eff057f57f1c02893522ad634898ab7d8" exitCode=0 Dec 03 06:57:23 crc kubenswrapper[4946]: I1203 06:57:23.297996 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"234eaea414d7f69a04537ca3a86c1a0eff057f57f1c02893522ad634898ab7d8"} Dec 03 06:57:23 crc kubenswrapper[4946]: I1203 06:57:23.298068 4946 scope.go:117] "RemoveContainer" containerID="5296d2535d9eb6b4c292ddcf0e7b560b6329e734738832cf71da39ae1ad35a45" Dec 03 06:57:24 crc kubenswrapper[4946]: I1203 06:57:24.306538 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"534a14b3a7d4edf5546e1374aec22c148fc1d6414e3e70ca3d3e3f20d48b2e02"} Dec 03 06:59:23 crc kubenswrapper[4946]: I1203 06:59:23.039365 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:59:23 crc kubenswrapper[4946]: I1203 06:59:23.040259 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:59:53 crc kubenswrapper[4946]: I1203 06:59:53.039312 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:59:53 crc kubenswrapper[4946]: I1203 06:59:53.039996 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.198505 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc"] Dec 03 07:00:00 crc kubenswrapper[4946]: E1203 07:00:00.199237 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b8c7afc-56fb-49ab-ae82-e60167809f15" containerName="registry" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.199256 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b8c7afc-56fb-49ab-ae82-e60167809f15" containerName="registry" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.199434 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b8c7afc-56fb-49ab-ae82-e60167809f15" containerName="registry" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.200042 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.202414 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.205598 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.214365 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc"] Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.308251 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b1cf6c8-022c-4f2b-b91f-4472579546cb-config-volume\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.308362 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b1cf6c8-022c-4f2b-b91f-4472579546cb-secret-volume\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.308415 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzgmq\" (UniqueName: \"kubernetes.io/projected/2b1cf6c8-022c-4f2b-b91f-4472579546cb-kube-api-access-jzgmq\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.409487 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b1cf6c8-022c-4f2b-b91f-4472579546cb-secret-volume\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.409553 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzgmq\" (UniqueName: \"kubernetes.io/projected/2b1cf6c8-022c-4f2b-b91f-4472579546cb-kube-api-access-jzgmq\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.409580 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b1cf6c8-022c-4f2b-b91f-4472579546cb-config-volume\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.410457 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b1cf6c8-022c-4f2b-b91f-4472579546cb-config-volume\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.417264 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b1cf6c8-022c-4f2b-b91f-4472579546cb-secret-volume\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.427927 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzgmq\" (UniqueName: \"kubernetes.io/projected/2b1cf6c8-022c-4f2b-b91f-4472579546cb-kube-api-access-jzgmq\") pod \"collect-profiles-29412420-zj9cc\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:00 crc kubenswrapper[4946]: I1203 07:00:00.520123 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:01 crc kubenswrapper[4946]: I1203 07:00:01.002374 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc"] Dec 03 07:00:01 crc kubenswrapper[4946]: I1203 07:00:01.288382 4946 generic.go:334] "Generic (PLEG): container finished" podID="2b1cf6c8-022c-4f2b-b91f-4472579546cb" containerID="5098e930dbe5b23b7e8b9f2804444be8d5fbeba493e0d2ffa76e15a93d2130fe" exitCode=0 Dec 03 07:00:01 crc kubenswrapper[4946]: I1203 07:00:01.288431 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" event={"ID":"2b1cf6c8-022c-4f2b-b91f-4472579546cb","Type":"ContainerDied","Data":"5098e930dbe5b23b7e8b9f2804444be8d5fbeba493e0d2ffa76e15a93d2130fe"} Dec 03 07:00:01 crc kubenswrapper[4946]: I1203 07:00:01.288772 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" event={"ID":"2b1cf6c8-022c-4f2b-b91f-4472579546cb","Type":"ContainerStarted","Data":"5ab838386ddf311ec40cbb171dde6dcea360441ef9252afa480a3f978fae25da"} Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.518708 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.637873 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b1cf6c8-022c-4f2b-b91f-4472579546cb-config-volume\") pod \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.638041 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzgmq\" (UniqueName: \"kubernetes.io/projected/2b1cf6c8-022c-4f2b-b91f-4472579546cb-kube-api-access-jzgmq\") pod \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.638115 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b1cf6c8-022c-4f2b-b91f-4472579546cb-secret-volume\") pod \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\" (UID: \"2b1cf6c8-022c-4f2b-b91f-4472579546cb\") " Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.638862 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1cf6c8-022c-4f2b-b91f-4472579546cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "2b1cf6c8-022c-4f2b-b91f-4472579546cb" (UID: "2b1cf6c8-022c-4f2b-b91f-4472579546cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.644414 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b1cf6c8-022c-4f2b-b91f-4472579546cb-kube-api-access-jzgmq" (OuterVolumeSpecName: "kube-api-access-jzgmq") pod "2b1cf6c8-022c-4f2b-b91f-4472579546cb" (UID: "2b1cf6c8-022c-4f2b-b91f-4472579546cb"). InnerVolumeSpecName "kube-api-access-jzgmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.644431 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b1cf6c8-022c-4f2b-b91f-4472579546cb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2b1cf6c8-022c-4f2b-b91f-4472579546cb" (UID: "2b1cf6c8-022c-4f2b-b91f-4472579546cb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.739935 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b1cf6c8-022c-4f2b-b91f-4472579546cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.740008 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzgmq\" (UniqueName: \"kubernetes.io/projected/2b1cf6c8-022c-4f2b-b91f-4472579546cb-kube-api-access-jzgmq\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:02 crc kubenswrapper[4946]: I1203 07:00:02.740033 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b1cf6c8-022c-4f2b-b91f-4472579546cb-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:03 crc kubenswrapper[4946]: I1203 07:00:03.301810 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" event={"ID":"2b1cf6c8-022c-4f2b-b91f-4472579546cb","Type":"ContainerDied","Data":"5ab838386ddf311ec40cbb171dde6dcea360441ef9252afa480a3f978fae25da"} Dec 03 07:00:03 crc kubenswrapper[4946]: I1203 07:00:03.301866 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ab838386ddf311ec40cbb171dde6dcea360441ef9252afa480a3f978fae25da" Dec 03 07:00:03 crc kubenswrapper[4946]: I1203 07:00:03.302267 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc" Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.039920 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.040532 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.040578 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.041169 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"534a14b3a7d4edf5546e1374aec22c148fc1d6414e3e70ca3d3e3f20d48b2e02"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.041243 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://534a14b3a7d4edf5546e1374aec22c148fc1d6414e3e70ca3d3e3f20d48b2e02" gracePeriod=600 Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.439506 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="534a14b3a7d4edf5546e1374aec22c148fc1d6414e3e70ca3d3e3f20d48b2e02" exitCode=0 Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.439558 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"534a14b3a7d4edf5546e1374aec22c148fc1d6414e3e70ca3d3e3f20d48b2e02"} Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.439946 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"5f5bf61ac3d07d18e68762b57ed405ab4ef0e6a554347d633d86406fb904be5c"} Dec 03 07:00:23 crc kubenswrapper[4946]: I1203 07:00:23.439978 4946 scope.go:117] "RemoveContainer" containerID="234eaea414d7f69a04537ca3a86c1a0eff057f57f1c02893522ad634898ab7d8" Dec 03 07:02:23 crc kubenswrapper[4946]: I1203 07:02:23.039094 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:02:23 crc kubenswrapper[4946]: I1203 07:02:23.039769 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:02:47 crc kubenswrapper[4946]: I1203 07:02:47.492974 4946 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 07:02:53 crc kubenswrapper[4946]: I1203 07:02:53.039990 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:02:53 crc kubenswrapper[4946]: I1203 07:02:53.040399 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.039527 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.040223 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.040295 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.041241 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f5bf61ac3d07d18e68762b57ed405ab4ef0e6a554347d633d86406fb904be5c"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.041357 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://5f5bf61ac3d07d18e68762b57ed405ab4ef0e6a554347d633d86406fb904be5c" gracePeriod=600 Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.157946 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p8fhs"] Dec 03 07:03:23 crc kubenswrapper[4946]: E1203 07:03:23.158338 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1cf6c8-022c-4f2b-b91f-4472579546cb" containerName="collect-profiles" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.158514 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1cf6c8-022c-4f2b-b91f-4472579546cb" containerName="collect-profiles" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.160711 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b1cf6c8-022c-4f2b-b91f-4472579546cb" containerName="collect-profiles" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.173435 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.196322 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8fhs"] Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.312522 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-utilities\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.312580 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7wwb\" (UniqueName: \"kubernetes.io/projected/18177911-a7ff-4007-8e41-a2cdd173c94e-kube-api-access-d7wwb\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.312661 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-catalog-content\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.414146 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-catalog-content\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.414210 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-utilities\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.414233 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7wwb\" (UniqueName: \"kubernetes.io/projected/18177911-a7ff-4007-8e41-a2cdd173c94e-kube-api-access-d7wwb\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.414618 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-catalog-content\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.414672 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-utilities\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.433049 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7wwb\" (UniqueName: \"kubernetes.io/projected/18177911-a7ff-4007-8e41-a2cdd173c94e-kube-api-access-d7wwb\") pod \"redhat-marketplace-p8fhs\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.510194 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:23 crc kubenswrapper[4946]: I1203 07:03:23.736080 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8fhs"] Dec 03 07:03:24 crc kubenswrapper[4946]: I1203 07:03:24.186330 4946 generic.go:334] "Generic (PLEG): container finished" podID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerID="eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740" exitCode=0 Dec 03 07:03:24 crc kubenswrapper[4946]: I1203 07:03:24.186438 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8fhs" event={"ID":"18177911-a7ff-4007-8e41-a2cdd173c94e","Type":"ContainerDied","Data":"eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740"} Dec 03 07:03:24 crc kubenswrapper[4946]: I1203 07:03:24.186823 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8fhs" event={"ID":"18177911-a7ff-4007-8e41-a2cdd173c94e","Type":"ContainerStarted","Data":"a4c15e074b6147a0f93b15e705e5589602f7cee5e1cf62910c7cc642e49834fe"} Dec 03 07:03:24 crc kubenswrapper[4946]: I1203 07:03:24.189158 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:03:24 crc kubenswrapper[4946]: I1203 07:03:24.198859 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="5f5bf61ac3d07d18e68762b57ed405ab4ef0e6a554347d633d86406fb904be5c" exitCode=0 Dec 03 07:03:24 crc kubenswrapper[4946]: I1203 07:03:24.198911 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"5f5bf61ac3d07d18e68762b57ed405ab4ef0e6a554347d633d86406fb904be5c"} Dec 03 07:03:24 crc kubenswrapper[4946]: I1203 07:03:24.198963 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"eb200c920ce9a9189cb9d22399d35826a7f0ee55cd4d5b661511822e8281f026"} Dec 03 07:03:24 crc kubenswrapper[4946]: I1203 07:03:24.198987 4946 scope.go:117] "RemoveContainer" containerID="534a14b3a7d4edf5546e1374aec22c148fc1d6414e3e70ca3d3e3f20d48b2e02" Dec 03 07:03:25 crc kubenswrapper[4946]: I1203 07:03:25.211348 4946 generic.go:334] "Generic (PLEG): container finished" podID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerID="a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a" exitCode=0 Dec 03 07:03:25 crc kubenswrapper[4946]: I1203 07:03:25.211447 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8fhs" event={"ID":"18177911-a7ff-4007-8e41-a2cdd173c94e","Type":"ContainerDied","Data":"a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a"} Dec 03 07:03:26 crc kubenswrapper[4946]: I1203 07:03:26.226593 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8fhs" event={"ID":"18177911-a7ff-4007-8e41-a2cdd173c94e","Type":"ContainerStarted","Data":"644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4"} Dec 03 07:03:26 crc kubenswrapper[4946]: I1203 07:03:26.254093 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p8fhs" podStartSLOduration=1.4015531270000001 podStartE2EDuration="3.254067294s" podCreationTimestamp="2025-12-03 07:03:23 +0000 UTC" firstStartedPulling="2025-12-03 07:03:24.188447007 +0000 UTC m=+796.985137136" lastFinishedPulling="2025-12-03 07:03:26.040961184 +0000 UTC m=+798.837651303" observedRunningTime="2025-12-03 07:03:26.248476265 +0000 UTC m=+799.045166384" watchObservedRunningTime="2025-12-03 07:03:26.254067294 +0000 UTC m=+799.050757443" Dec 03 07:03:33 crc kubenswrapper[4946]: I1203 07:03:33.510733 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:33 crc kubenswrapper[4946]: I1203 07:03:33.511451 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:33 crc kubenswrapper[4946]: I1203 07:03:33.574924 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:34 crc kubenswrapper[4946]: I1203 07:03:34.348396 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:34 crc kubenswrapper[4946]: I1203 07:03:34.409550 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8fhs"] Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.286052 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p8fhs" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerName="registry-server" containerID="cri-o://644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4" gracePeriod=2 Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.690644 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.890143 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-catalog-content\") pod \"18177911-a7ff-4007-8e41-a2cdd173c94e\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.890203 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-utilities\") pod \"18177911-a7ff-4007-8e41-a2cdd173c94e\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.890237 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7wwb\" (UniqueName: \"kubernetes.io/projected/18177911-a7ff-4007-8e41-a2cdd173c94e-kube-api-access-d7wwb\") pod \"18177911-a7ff-4007-8e41-a2cdd173c94e\" (UID: \"18177911-a7ff-4007-8e41-a2cdd173c94e\") " Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.892542 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-utilities" (OuterVolumeSpecName: "utilities") pod "18177911-a7ff-4007-8e41-a2cdd173c94e" (UID: "18177911-a7ff-4007-8e41-a2cdd173c94e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.897274 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18177911-a7ff-4007-8e41-a2cdd173c94e-kube-api-access-d7wwb" (OuterVolumeSpecName: "kube-api-access-d7wwb") pod "18177911-a7ff-4007-8e41-a2cdd173c94e" (UID: "18177911-a7ff-4007-8e41-a2cdd173c94e"). InnerVolumeSpecName "kube-api-access-d7wwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.930993 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18177911-a7ff-4007-8e41-a2cdd173c94e" (UID: "18177911-a7ff-4007-8e41-a2cdd173c94e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.992151 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.992200 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18177911-a7ff-4007-8e41-a2cdd173c94e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:36 crc kubenswrapper[4946]: I1203 07:03:36.992220 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7wwb\" (UniqueName: \"kubernetes.io/projected/18177911-a7ff-4007-8e41-a2cdd173c94e-kube-api-access-d7wwb\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.296048 4946 generic.go:334] "Generic (PLEG): container finished" podID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerID="644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4" exitCode=0 Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.296094 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8fhs" event={"ID":"18177911-a7ff-4007-8e41-a2cdd173c94e","Type":"ContainerDied","Data":"644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4"} Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.296125 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8fhs" event={"ID":"18177911-a7ff-4007-8e41-a2cdd173c94e","Type":"ContainerDied","Data":"a4c15e074b6147a0f93b15e705e5589602f7cee5e1cf62910c7cc642e49834fe"} Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.296144 4946 scope.go:117] "RemoveContainer" containerID="644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.296288 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8fhs" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.330198 4946 scope.go:117] "RemoveContainer" containerID="a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.338620 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8fhs"] Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.344359 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8fhs"] Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.368464 4946 scope.go:117] "RemoveContainer" containerID="eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.383199 4946 scope.go:117] "RemoveContainer" containerID="644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4" Dec 03 07:03:37 crc kubenswrapper[4946]: E1203 07:03:37.384034 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4\": container with ID starting with 644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4 not found: ID does not exist" containerID="644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.384089 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4"} err="failed to get container status \"644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4\": rpc error: code = NotFound desc = could not find container \"644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4\": container with ID starting with 644302b34d3cf631848ebf10d2303eb7aff41ea4257b30cb84278f917358c2c4 not found: ID does not exist" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.384125 4946 scope.go:117] "RemoveContainer" containerID="a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a" Dec 03 07:03:37 crc kubenswrapper[4946]: E1203 07:03:37.384687 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a\": container with ID starting with a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a not found: ID does not exist" containerID="a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.384713 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a"} err="failed to get container status \"a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a\": rpc error: code = NotFound desc = could not find container \"a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a\": container with ID starting with a03e4f6527a009f31a003daa43e6e7de486f2e508d455350ed7c0aaa0009c96a not found: ID does not exist" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.384730 4946 scope.go:117] "RemoveContainer" containerID="eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740" Dec 03 07:03:37 crc kubenswrapper[4946]: E1203 07:03:37.385126 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740\": container with ID starting with eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740 not found: ID does not exist" containerID="eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.385176 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740"} err="failed to get container status \"eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740\": rpc error: code = NotFound desc = could not find container \"eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740\": container with ID starting with eedd39f9669c93210841550384c141043c90736426a613727aca62facd1f8740 not found: ID does not exist" Dec 03 07:03:37 crc kubenswrapper[4946]: I1203 07:03:37.599215 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" path="/var/lib/kubelet/pods/18177911-a7ff-4007-8e41-a2cdd173c94e/volumes" Dec 03 07:04:43 crc kubenswrapper[4946]: I1203 07:04:43.998887 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6fppr"] Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:43.999844 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovn-controller" containerID="cri-o://bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1" gracePeriod=30 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.000249 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="sbdb" containerID="cri-o://6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2" gracePeriod=30 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.000297 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="nbdb" containerID="cri-o://0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b" gracePeriod=30 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.000340 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="northd" containerID="cri-o://2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb" gracePeriod=30 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.000416 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18" gracePeriod=30 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.000469 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kube-rbac-proxy-node" containerID="cri-o://9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0" gracePeriod=30 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.000511 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovn-acl-logging" containerID="cri-o://8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de" gracePeriod=30 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.038102 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" containerID="cri-o://7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" gracePeriod=30 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.338130 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/3.log" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.340551 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovn-acl-logging/0.log" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.341101 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovn-controller/0.log" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.341914 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.404961 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-h7zhq"] Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405235 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405263 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405278 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405290 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405303 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405316 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405329 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerName="extract-utilities" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405339 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerName="extract-utilities" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405354 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerName="registry-server" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405365 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerName="registry-server" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405378 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405389 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405401 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405412 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405427 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kube-rbac-proxy-node" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405437 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kube-rbac-proxy-node" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405454 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405465 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405479 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovn-acl-logging" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405489 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovn-acl-logging" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405503 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="northd" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405515 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="northd" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405532 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="sbdb" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405540 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="sbdb" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405551 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kubecfg-setup" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405560 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kubecfg-setup" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405571 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovn-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405579 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovn-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405589 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerName="extract-content" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405597 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerName="extract-content" Dec 03 07:04:44 crc kubenswrapper[4946]: E1203 07:04:44.405613 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="nbdb" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405621 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="nbdb" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405771 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovn-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405784 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovn-acl-logging" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405796 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405805 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405817 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405828 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405838 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405853 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="nbdb" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405863 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="18177911-a7ff-4007-8e41-a2cdd173c94e" containerName="registry-server" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405872 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="northd" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405961 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="kube-rbac-proxy-node" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.405973 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="sbdb" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.406187 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerName="ovnkube-controller" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.408194 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472403 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-config\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472472 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-systemd\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472514 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzcsn\" (UniqueName: \"kubernetes.io/projected/58ed82b9-82e5-4d9a-b331-80c32397dc43-kube-api-access-lzcsn\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472552 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovn-node-metrics-cert\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472584 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-slash\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472623 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-script-lib\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472654 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-netd\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472701 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-var-lib-openvswitch\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472869 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-etc-openvswitch\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472921 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-ovn\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472950 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-ovn-kubernetes\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.472978 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-node-log\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473027 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-systemd-units\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473034 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473086 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-kubelet\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473103 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473122 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-netns\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473161 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-env-overrides\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473204 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-openvswitch\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473235 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-var-lib-cni-networks-ovn-kubernetes\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473274 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-bin\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473322 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-log-socket\") pod \"58ed82b9-82e5-4d9a-b331-80c32397dc43\" (UID: \"58ed82b9-82e5-4d9a-b331-80c32397dc43\") " Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.473499 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-slash" (OuterVolumeSpecName: "host-slash") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474039 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474040 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474092 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474129 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474140 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474165 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474172 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474183 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474201 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-node-log" (OuterVolumeSpecName: "node-log") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474215 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474255 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-log-socket" (OuterVolumeSpecName: "log-socket") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474294 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474535 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474626 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474967 4946 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.474991 4946 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475005 4946 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475020 4946 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475032 4946 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475045 4946 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475056 4946 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475068 4946 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475079 4946 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475090 4946 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475101 4946 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475113 4946 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475124 4946 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475136 4946 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475147 4946 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475157 4946 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.475168 4946 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/58ed82b9-82e5-4d9a-b331-80c32397dc43-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.479305 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58ed82b9-82e5-4d9a-b331-80c32397dc43-kube-api-access-lzcsn" (OuterVolumeSpecName: "kube-api-access-lzcsn") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "kube-api-access-lzcsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.479852 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.495717 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "58ed82b9-82e5-4d9a-b331-80c32397dc43" (UID: "58ed82b9-82e5-4d9a-b331-80c32397dc43"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576217 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-slash\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576262 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovn-node-metrics-cert\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576281 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576297 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74hfv\" (UniqueName: \"kubernetes.io/projected/7464a7bc-6b3e-4d53-a185-ae41a8fece28-kube-api-access-74hfv\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576317 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-node-log\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576334 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-run-ovn-kubernetes\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576350 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovnkube-script-lib\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576485 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-cni-netd\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576503 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-kubelet\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576522 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-run-netns\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576534 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-log-socket\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576554 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-systemd\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576569 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-env-overrides\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576584 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-etc-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576602 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-systemd-units\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576641 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-ovn\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.576846 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-var-lib-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.577100 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovnkube-config\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.577211 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.577283 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-cni-bin\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.577490 4946 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/58ed82b9-82e5-4d9a-b331-80c32397dc43-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.577522 4946 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/58ed82b9-82e5-4d9a-b331-80c32397dc43-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.577550 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzcsn\" (UniqueName: \"kubernetes.io/projected/58ed82b9-82e5-4d9a-b331-80c32397dc43-kube-api-access-lzcsn\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678363 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-run-ovn-kubernetes\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678453 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovnkube-script-lib\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678527 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-cni-netd\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678530 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-run-ovn-kubernetes\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678570 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-kubelet\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678705 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-kubelet\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678757 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-run-netns\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678795 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-log-socket\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678839 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-systemd\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678874 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-env-overrides\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678905 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-etc-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678913 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-systemd\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678941 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-systemd-units\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678962 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-run-netns\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.678995 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-etc-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679008 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-ovn\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679045 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-systemd-units\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679045 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-ovn\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679082 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-var-lib-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679133 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovnkube-config\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679169 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-var-lib-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679181 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679223 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-cni-bin\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679263 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovnkube-script-lib\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679272 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679270 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-slash\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679310 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-slash\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679332 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-cni-bin\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679339 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovn-node-metrics-cert\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679387 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679420 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74hfv\" (UniqueName: \"kubernetes.io/projected/7464a7bc-6b3e-4d53-a185-ae41a8fece28-kube-api-access-74hfv\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679456 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-node-log\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679463 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-run-openvswitch\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.679536 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-node-log\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.680010 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-env-overrides\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.680124 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-log-socket\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.680183 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7464a7bc-6b3e-4d53-a185-ae41a8fece28-host-cni-netd\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.680482 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovnkube-config\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.685187 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7464a7bc-6b3e-4d53-a185-ae41a8fece28-ovn-node-metrics-cert\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.707283 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74hfv\" (UniqueName: \"kubernetes.io/projected/7464a7bc-6b3e-4d53-a185-ae41a8fece28-kube-api-access-74hfv\") pod \"ovnkube-node-h7zhq\" (UID: \"7464a7bc-6b3e-4d53-a185-ae41a8fece28\") " pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.724171 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.729881 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovnkube-controller/3.log" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.733662 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovn-acl-logging/0.log" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.734342 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6fppr_58ed82b9-82e5-4d9a-b331-80c32397dc43/ovn-controller/0.log" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.734994 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" exitCode=0 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735049 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2" exitCode=0 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735071 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b" exitCode=0 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735095 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb" exitCode=0 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735114 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18" exitCode=0 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735131 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0" exitCode=0 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735149 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de" exitCode=143 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735167 4946 generic.go:334] "Generic (PLEG): container finished" podID="58ed82b9-82e5-4d9a-b331-80c32397dc43" containerID="bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1" exitCode=143 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735247 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735326 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735330 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735350 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735370 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735422 4946 scope.go:117] "RemoveContainer" containerID="7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.735600 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736460 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736548 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736571 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736583 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736594 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736605 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736616 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736626 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736637 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736647 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736663 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736680 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736693 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736707 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736719 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736730 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736769 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736785 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736799 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736813 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736826 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736886 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736916 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736932 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736946 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736957 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736968 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736980 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.736990 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737002 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737013 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737024 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737040 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6fppr" event={"ID":"58ed82b9-82e5-4d9a-b331-80c32397dc43","Type":"ContainerDied","Data":"25638cd36b4f7f640ad6d77aa55eeb01b44bddc032d173b438aa5bd72d527b63"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737057 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737071 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737082 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737092 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737104 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737114 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737125 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737136 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737146 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.737158 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.740954 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/2.log" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.742362 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/1.log" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.742442 4946 generic.go:334] "Generic (PLEG): container finished" podID="08074f18-fe84-4d7b-8327-9696cbe78f38" containerID="7570afed4df875e4bd9bb2d80ac89f61f8a0151c58f9912a40f360f2d41528d2" exitCode=2 Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.742490 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t9hvz" event={"ID":"08074f18-fe84-4d7b-8327-9696cbe78f38","Type":"ContainerDied","Data":"7570afed4df875e4bd9bb2d80ac89f61f8a0151c58f9912a40f360f2d41528d2"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.742529 4946 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59"} Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.743225 4946 scope.go:117] "RemoveContainer" containerID="7570afed4df875e4bd9bb2d80ac89f61f8a0151c58f9912a40f360f2d41528d2" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.775972 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.811959 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6fppr"] Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.817985 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6fppr"] Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.827046 4946 scope.go:117] "RemoveContainer" containerID="6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.859354 4946 scope.go:117] "RemoveContainer" containerID="0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.875288 4946 scope.go:117] "RemoveContainer" containerID="2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.895498 4946 scope.go:117] "RemoveContainer" containerID="47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18" Dec 03 07:04:44 crc kubenswrapper[4946]: I1203 07:04:44.917191 4946 scope.go:117] "RemoveContainer" containerID="9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:44.999565 4946 scope.go:117] "RemoveContainer" containerID="8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.012824 4946 scope.go:117] "RemoveContainer" containerID="bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.026333 4946 scope.go:117] "RemoveContainer" containerID="61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.042883 4946 scope.go:117] "RemoveContainer" containerID="7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.043556 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": container with ID starting with 7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4 not found: ID does not exist" containerID="7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.043600 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} err="failed to get container status \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": rpc error: code = NotFound desc = could not find container \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": container with ID starting with 7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.043627 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.043983 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": container with ID starting with 6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816 not found: ID does not exist" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.044008 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} err="failed to get container status \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": rpc error: code = NotFound desc = could not find container \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": container with ID starting with 6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.044023 4946 scope.go:117] "RemoveContainer" containerID="6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.044310 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": container with ID starting with 6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2 not found: ID does not exist" containerID="6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.044367 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} err="failed to get container status \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": rpc error: code = NotFound desc = could not find container \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": container with ID starting with 6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.044441 4946 scope.go:117] "RemoveContainer" containerID="0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.044796 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": container with ID starting with 0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b not found: ID does not exist" containerID="0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.044819 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} err="failed to get container status \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": rpc error: code = NotFound desc = could not find container \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": container with ID starting with 0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.044834 4946 scope.go:117] "RemoveContainer" containerID="2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.045108 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": container with ID starting with 2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb not found: ID does not exist" containerID="2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.045151 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} err="failed to get container status \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": rpc error: code = NotFound desc = could not find container \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": container with ID starting with 2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.045179 4946 scope.go:117] "RemoveContainer" containerID="47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.045623 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": container with ID starting with 47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18 not found: ID does not exist" containerID="47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.045647 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} err="failed to get container status \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": rpc error: code = NotFound desc = could not find container \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": container with ID starting with 47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.045664 4946 scope.go:117] "RemoveContainer" containerID="9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.046064 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": container with ID starting with 9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0 not found: ID does not exist" containerID="9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.046097 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} err="failed to get container status \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": rpc error: code = NotFound desc = could not find container \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": container with ID starting with 9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.046119 4946 scope.go:117] "RemoveContainer" containerID="8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.046402 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": container with ID starting with 8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de not found: ID does not exist" containerID="8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.046433 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} err="failed to get container status \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": rpc error: code = NotFound desc = could not find container \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": container with ID starting with 8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.046451 4946 scope.go:117] "RemoveContainer" containerID="bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.046990 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": container with ID starting with bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1 not found: ID does not exist" containerID="bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.047020 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} err="failed to get container status \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": rpc error: code = NotFound desc = could not find container \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": container with ID starting with bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.047039 4946 scope.go:117] "RemoveContainer" containerID="61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c" Dec 03 07:04:45 crc kubenswrapper[4946]: E1203 07:04:45.047294 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": container with ID starting with 61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c not found: ID does not exist" containerID="61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.047321 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} err="failed to get container status \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": rpc error: code = NotFound desc = could not find container \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": container with ID starting with 61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.047336 4946 scope.go:117] "RemoveContainer" containerID="7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.047576 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} err="failed to get container status \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": rpc error: code = NotFound desc = could not find container \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": container with ID starting with 7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.047607 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.047873 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} err="failed to get container status \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": rpc error: code = NotFound desc = could not find container \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": container with ID starting with 6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.047898 4946 scope.go:117] "RemoveContainer" containerID="6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.048130 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} err="failed to get container status \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": rpc error: code = NotFound desc = could not find container \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": container with ID starting with 6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.048153 4946 scope.go:117] "RemoveContainer" containerID="0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.048414 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} err="failed to get container status \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": rpc error: code = NotFound desc = could not find container \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": container with ID starting with 0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.048437 4946 scope.go:117] "RemoveContainer" containerID="2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.048658 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} err="failed to get container status \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": rpc error: code = NotFound desc = could not find container \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": container with ID starting with 2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.048683 4946 scope.go:117] "RemoveContainer" containerID="47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.049110 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} err="failed to get container status \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": rpc error: code = NotFound desc = could not find container \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": container with ID starting with 47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.049135 4946 scope.go:117] "RemoveContainer" containerID="9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.049414 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} err="failed to get container status \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": rpc error: code = NotFound desc = could not find container \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": container with ID starting with 9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.049438 4946 scope.go:117] "RemoveContainer" containerID="8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.049691 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} err="failed to get container status \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": rpc error: code = NotFound desc = could not find container \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": container with ID starting with 8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.049719 4946 scope.go:117] "RemoveContainer" containerID="bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.050029 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} err="failed to get container status \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": rpc error: code = NotFound desc = could not find container \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": container with ID starting with bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.050056 4946 scope.go:117] "RemoveContainer" containerID="61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.050326 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} err="failed to get container status \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": rpc error: code = NotFound desc = could not find container \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": container with ID starting with 61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.050350 4946 scope.go:117] "RemoveContainer" containerID="7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.051218 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} err="failed to get container status \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": rpc error: code = NotFound desc = could not find container \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": container with ID starting with 7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.051257 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.051614 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} err="failed to get container status \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": rpc error: code = NotFound desc = could not find container \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": container with ID starting with 6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.051642 4946 scope.go:117] "RemoveContainer" containerID="6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.052203 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} err="failed to get container status \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": rpc error: code = NotFound desc = could not find container \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": container with ID starting with 6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.052230 4946 scope.go:117] "RemoveContainer" containerID="0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.052564 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} err="failed to get container status \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": rpc error: code = NotFound desc = could not find container \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": container with ID starting with 0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.052598 4946 scope.go:117] "RemoveContainer" containerID="2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.053175 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} err="failed to get container status \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": rpc error: code = NotFound desc = could not find container \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": container with ID starting with 2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.053202 4946 scope.go:117] "RemoveContainer" containerID="47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.053457 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} err="failed to get container status \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": rpc error: code = NotFound desc = could not find container \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": container with ID starting with 47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.053490 4946 scope.go:117] "RemoveContainer" containerID="9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.053729 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} err="failed to get container status \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": rpc error: code = NotFound desc = could not find container \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": container with ID starting with 9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.053767 4946 scope.go:117] "RemoveContainer" containerID="8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.054245 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} err="failed to get container status \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": rpc error: code = NotFound desc = could not find container \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": container with ID starting with 8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.054276 4946 scope.go:117] "RemoveContainer" containerID="bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.054608 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} err="failed to get container status \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": rpc error: code = NotFound desc = could not find container \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": container with ID starting with bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.054649 4946 scope.go:117] "RemoveContainer" containerID="61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.054977 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} err="failed to get container status \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": rpc error: code = NotFound desc = could not find container \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": container with ID starting with 61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.055000 4946 scope.go:117] "RemoveContainer" containerID="7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.055288 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} err="failed to get container status \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": rpc error: code = NotFound desc = could not find container \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": container with ID starting with 7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.055318 4946 scope.go:117] "RemoveContainer" containerID="6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.055582 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816"} err="failed to get container status \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": rpc error: code = NotFound desc = could not find container \"6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816\": container with ID starting with 6a05f79371e958f187a93daab824c39ad35b77e4f7b4c5826a55f598b5155816 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.055610 4946 scope.go:117] "RemoveContainer" containerID="6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.056300 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2"} err="failed to get container status \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": rpc error: code = NotFound desc = could not find container \"6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2\": container with ID starting with 6a102e962b4985cfa7e090d1e954f226b46e6a52592bccba203bb6b9f3b601e2 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.056329 4946 scope.go:117] "RemoveContainer" containerID="0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.056776 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b"} err="failed to get container status \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": rpc error: code = NotFound desc = could not find container \"0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b\": container with ID starting with 0c21411fa64d9ad733b8fb14bd7e07098a544cd32744eb4bf42be40710ddab1b not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.056797 4946 scope.go:117] "RemoveContainer" containerID="2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.057158 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb"} err="failed to get container status \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": rpc error: code = NotFound desc = could not find container \"2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb\": container with ID starting with 2aae5fbd585bdcb1a2714804ac38eeddfd082f438dc215f6d0c63e88a82decfb not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.057180 4946 scope.go:117] "RemoveContainer" containerID="47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.057413 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18"} err="failed to get container status \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": rpc error: code = NotFound desc = could not find container \"47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18\": container with ID starting with 47bcfb356c5e328bc0586211ccc7bd38248aa9b90d675ffcadfd327f4f958f18 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.057452 4946 scope.go:117] "RemoveContainer" containerID="9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.057713 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0"} err="failed to get container status \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": rpc error: code = NotFound desc = could not find container \"9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0\": container with ID starting with 9b2e1c55d35681bdf99eb0275f2d5e03a3805d8180cee126c9c71ea1b2ed18a0 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.057749 4946 scope.go:117] "RemoveContainer" containerID="8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.058034 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de"} err="failed to get container status \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": rpc error: code = NotFound desc = could not find container \"8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de\": container with ID starting with 8f5c266fc9fd9caa74f43d5c61c51ec95f925ea4dc7d4cc9a0e25433203038de not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.058069 4946 scope.go:117] "RemoveContainer" containerID="bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.058378 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1"} err="failed to get container status \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": rpc error: code = NotFound desc = could not find container \"bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1\": container with ID starting with bb94d7480d2bab7d4be833423bac591cbe92fdd1812f3f130f7c48f5e7be0bb1 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.058404 4946 scope.go:117] "RemoveContainer" containerID="61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.059039 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c"} err="failed to get container status \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": rpc error: code = NotFound desc = could not find container \"61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c\": container with ID starting with 61316be7da5097b296e62ab7ddc599c5e3dc6226b6f86a4d56daff2b0554a81c not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.059072 4946 scope.go:117] "RemoveContainer" containerID="7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.059319 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4"} err="failed to get container status \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": rpc error: code = NotFound desc = could not find container \"7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4\": container with ID starting with 7f412e406c84929b4eb83e0d1fe70d49c66cb3a292bf75c7fee4cc1d4bb1b3f4 not found: ID does not exist" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.601288 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58ed82b9-82e5-4d9a-b331-80c32397dc43" path="/var/lib/kubelet/pods/58ed82b9-82e5-4d9a-b331-80c32397dc43/volumes" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.751821 4946 generic.go:334] "Generic (PLEG): container finished" podID="7464a7bc-6b3e-4d53-a185-ae41a8fece28" containerID="0961a7bbc93f972630a9b7b7b399c79499304ec29d017337302a19a26cea3573" exitCode=0 Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.751905 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerDied","Data":"0961a7bbc93f972630a9b7b7b399c79499304ec29d017337302a19a26cea3573"} Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.751976 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"984b5016f6b38a499b6f0aede13e8e8e93186ea2c1e809d8017bfac0401ce85f"} Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.755056 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/2.log" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.755496 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/1.log" Dec 03 07:04:45 crc kubenswrapper[4946]: I1203 07:04:45.755540 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-t9hvz" event={"ID":"08074f18-fe84-4d7b-8327-9696cbe78f38","Type":"ContainerStarted","Data":"8c2e740a6f05dbd0ac3ca3f19abf0efe70102cd535733a3e19f4eda8064882b0"} Dec 03 07:04:46 crc kubenswrapper[4946]: I1203 07:04:46.764055 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"3153966bee219e62f820a745692237f5dbd469911964622c14cd2e4857fb3a62"} Dec 03 07:04:46 crc kubenswrapper[4946]: I1203 07:04:46.764662 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"c6f17ea3721a5e8cd0b66fb424c9ff0d5f6ada64889025640f003633582d6093"} Dec 03 07:04:46 crc kubenswrapper[4946]: I1203 07:04:46.764678 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"92d16ada3ba8a32d65e6ecf5d19433243089007b9567565f9e177ed4cea174d6"} Dec 03 07:04:46 crc kubenswrapper[4946]: I1203 07:04:46.764688 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"8ac8401ad2f2772c2e84b852d382ce3596f671ad7d45f8cef262368640e16a82"} Dec 03 07:04:46 crc kubenswrapper[4946]: I1203 07:04:46.764699 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"cba4fe36ffba56c88be3dc3b559cda135831f4f46d140a001213a704f4edf705"} Dec 03 07:04:46 crc kubenswrapper[4946]: I1203 07:04:46.764710 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"390d2e6da65fdde2617838698b06866a52798c8c2cac0b4f5e3848e87dc094c9"} Dec 03 07:04:49 crc kubenswrapper[4946]: I1203 07:04:49.781378 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"62b446d6cc0a048938adc33b8e547c8161ff06a880a970382535e887c1f94c8c"} Dec 03 07:04:52 crc kubenswrapper[4946]: I1203 07:04:52.800120 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" event={"ID":"7464a7bc-6b3e-4d53-a185-ae41a8fece28","Type":"ContainerStarted","Data":"501169c8dba95a8971e136bfac075588fc9568c0fc0d74e0d4a7d59467d545b4"} Dec 03 07:04:52 crc kubenswrapper[4946]: I1203 07:04:52.800583 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:52 crc kubenswrapper[4946]: I1203 07:04:52.800600 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:52 crc kubenswrapper[4946]: I1203 07:04:52.800610 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:52 crc kubenswrapper[4946]: I1203 07:04:52.825049 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:52 crc kubenswrapper[4946]: I1203 07:04:52.827172 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" podStartSLOduration=8.82714991 podStartE2EDuration="8.82714991s" podCreationTimestamp="2025-12-03 07:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:04:52.824166921 +0000 UTC m=+885.620857030" watchObservedRunningTime="2025-12-03 07:04:52.82714991 +0000 UTC m=+885.623840029" Dec 03 07:04:52 crc kubenswrapper[4946]: I1203 07:04:52.831514 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.771019 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-zklc4"] Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.772017 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.773873 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.774037 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.774404 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.774485 4946 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-gd7jn" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.781164 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-zklc4"] Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.836602 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-crc-storage\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.836654 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-node-mnt\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.836835 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hmpj\" (UniqueName: \"kubernetes.io/projected/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-kube-api-access-8hmpj\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.937962 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hmpj\" (UniqueName: \"kubernetes.io/projected/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-kube-api-access-8hmpj\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.938062 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-crc-storage\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.938136 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-node-mnt\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.938809 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-node-mnt\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.940124 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-crc-storage\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:56 crc kubenswrapper[4946]: I1203 07:04:56.958808 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hmpj\" (UniqueName: \"kubernetes.io/projected/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-kube-api-access-8hmpj\") pod \"crc-storage-crc-zklc4\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:57 crc kubenswrapper[4946]: I1203 07:04:57.092855 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:04:57 crc kubenswrapper[4946]: I1203 07:04:57.302617 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-zklc4"] Dec 03 07:04:57 crc kubenswrapper[4946]: W1203 07:04:57.306350 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd933dfaf_1545_4e02_8bf6_67b6be0cfd75.slice/crio-fe8c8593703617704f8eb2af0e2b7c5c35db002f8fdf3e2901b188e97e944347 WatchSource:0}: Error finding container fe8c8593703617704f8eb2af0e2b7c5c35db002f8fdf3e2901b188e97e944347: Status 404 returned error can't find the container with id fe8c8593703617704f8eb2af0e2b7c5c35db002f8fdf3e2901b188e97e944347 Dec 03 07:04:57 crc kubenswrapper[4946]: I1203 07:04:57.831274 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-zklc4" event={"ID":"d933dfaf-1545-4e02-8bf6-67b6be0cfd75","Type":"ContainerStarted","Data":"fe8c8593703617704f8eb2af0e2b7c5c35db002f8fdf3e2901b188e97e944347"} Dec 03 07:04:59 crc kubenswrapper[4946]: I1203 07:04:59.843448 4946 generic.go:334] "Generic (PLEG): container finished" podID="d933dfaf-1545-4e02-8bf6-67b6be0cfd75" containerID="6adf27e173781e9ff9424bfe807b466690c5219a30eb5239139d30e4068600ae" exitCode=0 Dec 03 07:04:59 crc kubenswrapper[4946]: I1203 07:04:59.843530 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-zklc4" event={"ID":"d933dfaf-1545-4e02-8bf6-67b6be0cfd75","Type":"ContainerDied","Data":"6adf27e173781e9ff9424bfe807b466690c5219a30eb5239139d30e4068600ae"} Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.175244 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.299400 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-crc-storage\") pod \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.299487 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-node-mnt\") pod \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.299614 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hmpj\" (UniqueName: \"kubernetes.io/projected/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-kube-api-access-8hmpj\") pod \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\" (UID: \"d933dfaf-1545-4e02-8bf6-67b6be0cfd75\") " Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.300071 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "d933dfaf-1545-4e02-8bf6-67b6be0cfd75" (UID: "d933dfaf-1545-4e02-8bf6-67b6be0cfd75"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.308043 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-kube-api-access-8hmpj" (OuterVolumeSpecName: "kube-api-access-8hmpj") pod "d933dfaf-1545-4e02-8bf6-67b6be0cfd75" (UID: "d933dfaf-1545-4e02-8bf6-67b6be0cfd75"). InnerVolumeSpecName "kube-api-access-8hmpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.318882 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "d933dfaf-1545-4e02-8bf6-67b6be0cfd75" (UID: "d933dfaf-1545-4e02-8bf6-67b6be0cfd75"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.401598 4946 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.401632 4946 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.401645 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hmpj\" (UniqueName: \"kubernetes.io/projected/d933dfaf-1545-4e02-8bf6-67b6be0cfd75-kube-api-access-8hmpj\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.861175 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-zklc4" event={"ID":"d933dfaf-1545-4e02-8bf6-67b6be0cfd75","Type":"ContainerDied","Data":"fe8c8593703617704f8eb2af0e2b7c5c35db002f8fdf3e2901b188e97e944347"} Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.861230 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe8c8593703617704f8eb2af0e2b7c5c35db002f8fdf3e2901b188e97e944347" Dec 03 07:05:01 crc kubenswrapper[4946]: I1203 07:05:01.861271 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-zklc4" Dec 03 07:05:07 crc kubenswrapper[4946]: I1203 07:05:07.909702 4946 scope.go:117] "RemoveContainer" containerID="e575197dfd684beb4c7a5a660d24f782573caf8048d064ae43c2a5a8e9100e59" Dec 03 07:05:08 crc kubenswrapper[4946]: I1203 07:05:08.918180 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-t9hvz_08074f18-fe84-4d7b-8327-9696cbe78f38/kube-multus/2.log" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.632967 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2"] Dec 03 07:05:09 crc kubenswrapper[4946]: E1203 07:05:09.633425 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d933dfaf-1545-4e02-8bf6-67b6be0cfd75" containerName="storage" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.633439 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d933dfaf-1545-4e02-8bf6-67b6be0cfd75" containerName="storage" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.633566 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d933dfaf-1545-4e02-8bf6-67b6be0cfd75" containerName="storage" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.634377 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.636962 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.645955 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2"] Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.810988 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.811102 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn7sj\" (UniqueName: \"kubernetes.io/projected/1e74163b-85c4-4470-b899-a40817bcfae9-kube-api-access-jn7sj\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.811492 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.912882 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn7sj\" (UniqueName: \"kubernetes.io/projected/1e74163b-85c4-4470-b899-a40817bcfae9-kube-api-access-jn7sj\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.913102 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.913168 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.914310 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.914360 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.949379 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn7sj\" (UniqueName: \"kubernetes.io/projected/1e74163b-85c4-4470-b899-a40817bcfae9-kube-api-access-jn7sj\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:09 crc kubenswrapper[4946]: I1203 07:05:09.958087 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:10 crc kubenswrapper[4946]: I1203 07:05:10.212953 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2"] Dec 03 07:05:10 crc kubenswrapper[4946]: I1203 07:05:10.935204 4946 generic.go:334] "Generic (PLEG): container finished" podID="1e74163b-85c4-4470-b899-a40817bcfae9" containerID="17e85070fb43899c4f2650f6902067e36a12d8fbdf1ee3d1935eadf0c7eb29fa" exitCode=0 Dec 03 07:05:10 crc kubenswrapper[4946]: I1203 07:05:10.935354 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" event={"ID":"1e74163b-85c4-4470-b899-a40817bcfae9","Type":"ContainerDied","Data":"17e85070fb43899c4f2650f6902067e36a12d8fbdf1ee3d1935eadf0c7eb29fa"} Dec 03 07:05:10 crc kubenswrapper[4946]: I1203 07:05:10.935999 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" event={"ID":"1e74163b-85c4-4470-b899-a40817bcfae9","Type":"ContainerStarted","Data":"21ab567c6e8331046f16a3a036cbf6fcce0bc678831c3055aeea79f6cf263f88"} Dec 03 07:05:11 crc kubenswrapper[4946]: I1203 07:05:11.935058 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k5vsn"] Dec 03 07:05:11 crc kubenswrapper[4946]: I1203 07:05:11.937872 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:11 crc kubenswrapper[4946]: I1203 07:05:11.953262 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k5vsn"] Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.041688 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-utilities\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.041966 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-catalog-content\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.042184 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-479vl\" (UniqueName: \"kubernetes.io/projected/cd72718c-9315-41b5-8f52-edfa443eb98f-kube-api-access-479vl\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.143632 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-utilities\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.143705 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-catalog-content\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.143765 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-479vl\" (UniqueName: \"kubernetes.io/projected/cd72718c-9315-41b5-8f52-edfa443eb98f-kube-api-access-479vl\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.144294 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-utilities\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.144596 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-catalog-content\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.164662 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-479vl\" (UniqueName: \"kubernetes.io/projected/cd72718c-9315-41b5-8f52-edfa443eb98f-kube-api-access-479vl\") pod \"redhat-operators-k5vsn\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.289871 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.476497 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k5vsn"] Dec 03 07:05:12 crc kubenswrapper[4946]: W1203 07:05:12.489573 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd72718c_9315_41b5_8f52_edfa443eb98f.slice/crio-84ab2691e2c22278deb1519bfe939a1376dc020ebe655ba862d57de6719ee263 WatchSource:0}: Error finding container 84ab2691e2c22278deb1519bfe939a1376dc020ebe655ba862d57de6719ee263: Status 404 returned error can't find the container with id 84ab2691e2c22278deb1519bfe939a1376dc020ebe655ba862d57de6719ee263 Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.958033 4946 generic.go:334] "Generic (PLEG): container finished" podID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerID="b684e8bd47f03c7bcbbd79d073124e78f7947f20c6a8313c5769f8bbc94d78d8" exitCode=0 Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.958115 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5vsn" event={"ID":"cd72718c-9315-41b5-8f52-edfa443eb98f","Type":"ContainerDied","Data":"b684e8bd47f03c7bcbbd79d073124e78f7947f20c6a8313c5769f8bbc94d78d8"} Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.958146 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5vsn" event={"ID":"cd72718c-9315-41b5-8f52-edfa443eb98f","Type":"ContainerStarted","Data":"84ab2691e2c22278deb1519bfe939a1376dc020ebe655ba862d57de6719ee263"} Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.961158 4946 generic.go:334] "Generic (PLEG): container finished" podID="1e74163b-85c4-4470-b899-a40817bcfae9" containerID="33cd9ca924949ae2a7253a334d3509b404fabfcc3996b1c2c42466aa9ce68a40" exitCode=0 Dec 03 07:05:12 crc kubenswrapper[4946]: I1203 07:05:12.961189 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" event={"ID":"1e74163b-85c4-4470-b899-a40817bcfae9","Type":"ContainerDied","Data":"33cd9ca924949ae2a7253a334d3509b404fabfcc3996b1c2c42466aa9ce68a40"} Dec 03 07:05:13 crc kubenswrapper[4946]: I1203 07:05:13.975414 4946 generic.go:334] "Generic (PLEG): container finished" podID="1e74163b-85c4-4470-b899-a40817bcfae9" containerID="3bd689daf8e6527e4ec541620852ddfe96458f76108453646f514c459e025c77" exitCode=0 Dec 03 07:05:13 crc kubenswrapper[4946]: I1203 07:05:13.975851 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" event={"ID":"1e74163b-85c4-4470-b899-a40817bcfae9","Type":"ContainerDied","Data":"3bd689daf8e6527e4ec541620852ddfe96458f76108453646f514c459e025c77"} Dec 03 07:05:14 crc kubenswrapper[4946]: I1203 07:05:14.760223 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h7zhq" Dec 03 07:05:14 crc kubenswrapper[4946]: I1203 07:05:14.984300 4946 generic.go:334] "Generic (PLEG): container finished" podID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerID="19685f4c4ddb5cdca9aff7c556c3c83c34d8b6fb47a59766e3cd35713df8578d" exitCode=0 Dec 03 07:05:14 crc kubenswrapper[4946]: I1203 07:05:14.984376 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5vsn" event={"ID":"cd72718c-9315-41b5-8f52-edfa443eb98f","Type":"ContainerDied","Data":"19685f4c4ddb5cdca9aff7c556c3c83c34d8b6fb47a59766e3cd35713df8578d"} Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.263923 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.385987 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn7sj\" (UniqueName: \"kubernetes.io/projected/1e74163b-85c4-4470-b899-a40817bcfae9-kube-api-access-jn7sj\") pod \"1e74163b-85c4-4470-b899-a40817bcfae9\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.386291 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-bundle\") pod \"1e74163b-85c4-4470-b899-a40817bcfae9\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.386316 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-util\") pod \"1e74163b-85c4-4470-b899-a40817bcfae9\" (UID: \"1e74163b-85c4-4470-b899-a40817bcfae9\") " Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.387577 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-bundle" (OuterVolumeSpecName: "bundle") pod "1e74163b-85c4-4470-b899-a40817bcfae9" (UID: "1e74163b-85c4-4470-b899-a40817bcfae9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.391315 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e74163b-85c4-4470-b899-a40817bcfae9-kube-api-access-jn7sj" (OuterVolumeSpecName: "kube-api-access-jn7sj") pod "1e74163b-85c4-4470-b899-a40817bcfae9" (UID: "1e74163b-85c4-4470-b899-a40817bcfae9"). InnerVolumeSpecName "kube-api-access-jn7sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.423972 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-util" (OuterVolumeSpecName: "util") pod "1e74163b-85c4-4470-b899-a40817bcfae9" (UID: "1e74163b-85c4-4470-b899-a40817bcfae9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.488384 4946 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.488442 4946 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e74163b-85c4-4470-b899-a40817bcfae9-util\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.488455 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn7sj\" (UniqueName: \"kubernetes.io/projected/1e74163b-85c4-4470-b899-a40817bcfae9-kube-api-access-jn7sj\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.995179 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5vsn" event={"ID":"cd72718c-9315-41b5-8f52-edfa443eb98f","Type":"ContainerStarted","Data":"617c377fb5733c3f58d40869c0146a64f08b8fd148cee24ee0a0b629a1871954"} Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.998795 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" event={"ID":"1e74163b-85c4-4470-b899-a40817bcfae9","Type":"ContainerDied","Data":"21ab567c6e8331046f16a3a036cbf6fcce0bc678831c3055aeea79f6cf263f88"} Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.998831 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21ab567c6e8331046f16a3a036cbf6fcce0bc678831c3055aeea79f6cf263f88" Dec 03 07:05:15 crc kubenswrapper[4946]: I1203 07:05:15.998889 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2" Dec 03 07:05:16 crc kubenswrapper[4946]: I1203 07:05:16.031133 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k5vsn" podStartSLOduration=2.534215957 podStartE2EDuration="5.03111111s" podCreationTimestamp="2025-12-03 07:05:11 +0000 UTC" firstStartedPulling="2025-12-03 07:05:12.959684204 +0000 UTC m=+905.756374313" lastFinishedPulling="2025-12-03 07:05:15.456579357 +0000 UTC m=+908.253269466" observedRunningTime="2025-12-03 07:05:16.025533021 +0000 UTC m=+908.822223200" watchObservedRunningTime="2025-12-03 07:05:16.03111111 +0000 UTC m=+908.827801249" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.231735 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt"] Dec 03 07:05:17 crc kubenswrapper[4946]: E1203 07:05:17.231986 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e74163b-85c4-4470-b899-a40817bcfae9" containerName="pull" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.232001 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e74163b-85c4-4470-b899-a40817bcfae9" containerName="pull" Dec 03 07:05:17 crc kubenswrapper[4946]: E1203 07:05:17.232021 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e74163b-85c4-4470-b899-a40817bcfae9" containerName="extract" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.232029 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e74163b-85c4-4470-b899-a40817bcfae9" containerName="extract" Dec 03 07:05:17 crc kubenswrapper[4946]: E1203 07:05:17.232046 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e74163b-85c4-4470-b899-a40817bcfae9" containerName="util" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.232054 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e74163b-85c4-4470-b899-a40817bcfae9" containerName="util" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.232164 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e74163b-85c4-4470-b899-a40817bcfae9" containerName="extract" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.232557 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.235233 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.235277 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-8d7x5" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.241122 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.250990 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt"] Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.415484 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhkrr\" (UniqueName: \"kubernetes.io/projected/0bcd8282-02a8-4389-bc9c-bb6d2b0904df-kube-api-access-fhkrr\") pod \"nmstate-operator-5b5b58f5c8-xjqdt\" (UID: \"0bcd8282-02a8-4389-bc9c-bb6d2b0904df\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.516582 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhkrr\" (UniqueName: \"kubernetes.io/projected/0bcd8282-02a8-4389-bc9c-bb6d2b0904df-kube-api-access-fhkrr\") pod \"nmstate-operator-5b5b58f5c8-xjqdt\" (UID: \"0bcd8282-02a8-4389-bc9c-bb6d2b0904df\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.536024 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhkrr\" (UniqueName: \"kubernetes.io/projected/0bcd8282-02a8-4389-bc9c-bb6d2b0904df-kube-api-access-fhkrr\") pod \"nmstate-operator-5b5b58f5c8-xjqdt\" (UID: \"0bcd8282-02a8-4389-bc9c-bb6d2b0904df\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.550785 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt" Dec 03 07:05:17 crc kubenswrapper[4946]: I1203 07:05:17.754479 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt"] Dec 03 07:05:18 crc kubenswrapper[4946]: I1203 07:05:18.008163 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt" event={"ID":"0bcd8282-02a8-4389-bc9c-bb6d2b0904df","Type":"ContainerStarted","Data":"4e198dc947d11445b5942972700db345df1aa77cce77166f9184a6d84a6ef7f5"} Dec 03 07:05:21 crc kubenswrapper[4946]: I1203 07:05:21.026026 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt" event={"ID":"0bcd8282-02a8-4389-bc9c-bb6d2b0904df","Type":"ContainerStarted","Data":"7e840dbfd56aea1bf6a9c7e1c7ca8e7f8affe4fabd2892e32e4c56f880234629"} Dec 03 07:05:21 crc kubenswrapper[4946]: I1203 07:05:21.048815 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xjqdt" podStartSLOduration=1.968353472 podStartE2EDuration="4.048793845s" podCreationTimestamp="2025-12-03 07:05:17 +0000 UTC" firstStartedPulling="2025-12-03 07:05:17.762279772 +0000 UTC m=+910.558969881" lastFinishedPulling="2025-12-03 07:05:19.842720155 +0000 UTC m=+912.639410254" observedRunningTime="2025-12-03 07:05:21.045277521 +0000 UTC m=+913.841967670" watchObservedRunningTime="2025-12-03 07:05:21.048793845 +0000 UTC m=+913.845483964" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.102656 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.103463 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.106190 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xf8n4" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.124419 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.135880 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.137106 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.140752 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.146799 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.173268 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vmz8j"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.174008 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.254472 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.255154 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.256838 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.257166 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.257332 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-w7w7g" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.275114 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.281798 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6rrb\" (UniqueName: \"kubernetes.io/projected/1e225c74-69b1-4094-9310-7e790d712264-kube-api-access-m6rrb\") pod \"nmstate-metrics-7f946cbc9-smzvj\" (UID: \"1e225c74-69b1-4094-9310-7e790d712264\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.281875 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-nmstate-lock\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.281932 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f636b80e-63df-4797-9720-bc7daa273204-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jx9l9\" (UID: \"f636b80e-63df-4797-9720-bc7daa273204\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.282000 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-ovs-socket\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.282039 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgfdx\" (UniqueName: \"kubernetes.io/projected/7219ceec-2cc3-4663-bf50-11ef855ec612-kube-api-access-bgfdx\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.282088 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqlt2\" (UniqueName: \"kubernetes.io/projected/f636b80e-63df-4797-9720-bc7daa273204-kube-api-access-pqlt2\") pod \"nmstate-webhook-5f6d4c5ccb-jx9l9\" (UID: \"f636b80e-63df-4797-9720-bc7daa273204\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.282119 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-dbus-socket\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.290995 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.291052 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.335180 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383279 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-ovs-socket\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383335 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgfdx\" (UniqueName: \"kubernetes.io/projected/7219ceec-2cc3-4663-bf50-11ef855ec612-kube-api-access-bgfdx\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383372 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ae0ab229-8baf-4c5b-a452-2fdd714946ea-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383407 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqlt2\" (UniqueName: \"kubernetes.io/projected/f636b80e-63df-4797-9720-bc7daa273204-kube-api-access-pqlt2\") pod \"nmstate-webhook-5f6d4c5ccb-jx9l9\" (UID: \"f636b80e-63df-4797-9720-bc7daa273204\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383436 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-dbus-socket\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383443 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-ovs-socket\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383481 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-nmstate-lock\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383507 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6rrb\" (UniqueName: \"kubernetes.io/projected/1e225c74-69b1-4094-9310-7e790d712264-kube-api-access-m6rrb\") pod \"nmstate-metrics-7f946cbc9-smzvj\" (UID: \"1e225c74-69b1-4094-9310-7e790d712264\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383538 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbxkt\" (UniqueName: \"kubernetes.io/projected/ae0ab229-8baf-4c5b-a452-2fdd714946ea-kube-api-access-bbxkt\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383573 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae0ab229-8baf-4c5b-a452-2fdd714946ea-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383599 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f636b80e-63df-4797-9720-bc7daa273204-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jx9l9\" (UID: \"f636b80e-63df-4797-9720-bc7daa273204\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.383767 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-nmstate-lock\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: E1203 07:05:22.383791 4946 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 03 07:05:22 crc kubenswrapper[4946]: E1203 07:05:22.383871 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f636b80e-63df-4797-9720-bc7daa273204-tls-key-pair podName:f636b80e-63df-4797-9720-bc7daa273204 nodeName:}" failed. No retries permitted until 2025-12-03 07:05:22.883826319 +0000 UTC m=+915.680516528 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/f636b80e-63df-4797-9720-bc7daa273204-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-jx9l9" (UID: "f636b80e-63df-4797-9720-bc7daa273204") : secret "openshift-nmstate-webhook" not found Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.384053 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7219ceec-2cc3-4663-bf50-11ef855ec612-dbus-socket\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.402539 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqlt2\" (UniqueName: \"kubernetes.io/projected/f636b80e-63df-4797-9720-bc7daa273204-kube-api-access-pqlt2\") pod \"nmstate-webhook-5f6d4c5ccb-jx9l9\" (UID: \"f636b80e-63df-4797-9720-bc7daa273204\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.405087 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6rrb\" (UniqueName: \"kubernetes.io/projected/1e225c74-69b1-4094-9310-7e790d712264-kube-api-access-m6rrb\") pod \"nmstate-metrics-7f946cbc9-smzvj\" (UID: \"1e225c74-69b1-4094-9310-7e790d712264\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.414365 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgfdx\" (UniqueName: \"kubernetes.io/projected/7219ceec-2cc3-4663-bf50-11ef855ec612-kube-api-access-bgfdx\") pod \"nmstate-handler-vmz8j\" (UID: \"7219ceec-2cc3-4663-bf50-11ef855ec612\") " pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.418797 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.480686 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6b5847d797-hxwsr"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.482982 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.484713 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbxkt\" (UniqueName: \"kubernetes.io/projected/ae0ab229-8baf-4c5b-a452-2fdd714946ea-kube-api-access-bbxkt\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.484789 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae0ab229-8baf-4c5b-a452-2fdd714946ea-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.484858 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ae0ab229-8baf-4c5b-a452-2fdd714946ea-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.485727 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ae0ab229-8baf-4c5b-a452-2fdd714946ea-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.491094 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b5847d797-hxwsr"] Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.497632 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.499975 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ae0ab229-8baf-4c5b-a452-2fdd714946ea-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.533603 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbxkt\" (UniqueName: \"kubernetes.io/projected/ae0ab229-8baf-4c5b-a452-2fdd714946ea-kube-api-access-bbxkt\") pod \"nmstate-console-plugin-7fbb5f6569-5jbpq\" (UID: \"ae0ab229-8baf-4c5b-a452-2fdd714946ea\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.574590 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.586889 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-oauth-serving-cert\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.586954 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-trusted-ca-bundle\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.587008 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z84x9\" (UniqueName: \"kubernetes.io/projected/741e35a4-2042-495b-853c-4b87a1f6d5bc-kube-api-access-z84x9\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.587098 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-config\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.587223 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-service-ca\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.587257 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-serving-cert\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.587494 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-oauth-config\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.688574 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-oauth-serving-cert\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.688631 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-trusted-ca-bundle\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.688672 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z84x9\" (UniqueName: \"kubernetes.io/projected/741e35a4-2042-495b-853c-4b87a1f6d5bc-kube-api-access-z84x9\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.688695 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-config\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.688724 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-service-ca\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.688771 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-serving-cert\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.688848 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-oauth-config\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.689791 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-config\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.690015 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-trusted-ca-bundle\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.689732 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-oauth-serving-cert\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.690784 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/741e35a4-2042-495b-853c-4b87a1f6d5bc-service-ca\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.693172 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-serving-cert\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.693520 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/741e35a4-2042-495b-853c-4b87a1f6d5bc-console-oauth-config\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.705521 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z84x9\" (UniqueName: \"kubernetes.io/projected/741e35a4-2042-495b-853c-4b87a1f6d5bc-kube-api-access-z84x9\") pod \"console-6b5847d797-hxwsr\" (UID: \"741e35a4-2042-495b-853c-4b87a1f6d5bc\") " pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.728102 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj"] Dec 03 07:05:22 crc kubenswrapper[4946]: W1203 07:05:22.735199 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e225c74_69b1_4094_9310_7e790d712264.slice/crio-0255a7d56477a1ddeaae4c370eac39e8dcb2b724237941aaec578d80f3cf49eb WatchSource:0}: Error finding container 0255a7d56477a1ddeaae4c370eac39e8dcb2b724237941aaec578d80f3cf49eb: Status 404 returned error can't find the container with id 0255a7d56477a1ddeaae4c370eac39e8dcb2b724237941aaec578d80f3cf49eb Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.779893 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq"] Dec 03 07:05:22 crc kubenswrapper[4946]: W1203 07:05:22.785917 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae0ab229_8baf_4c5b_a452_2fdd714946ea.slice/crio-241fbde969d71ef2c5c267eb26d0b60189bf3f71d3394a529b4a71e42c270e44 WatchSource:0}: Error finding container 241fbde969d71ef2c5c267eb26d0b60189bf3f71d3394a529b4a71e42c270e44: Status 404 returned error can't find the container with id 241fbde969d71ef2c5c267eb26d0b60189bf3f71d3394a529b4a71e42c270e44 Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.879655 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.891033 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f636b80e-63df-4797-9720-bc7daa273204-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jx9l9\" (UID: \"f636b80e-63df-4797-9720-bc7daa273204\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:22 crc kubenswrapper[4946]: I1203 07:05:22.895460 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f636b80e-63df-4797-9720-bc7daa273204-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jx9l9\" (UID: \"f636b80e-63df-4797-9720-bc7daa273204\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.037149 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" event={"ID":"1e225c74-69b1-4094-9310-7e790d712264","Type":"ContainerStarted","Data":"0255a7d56477a1ddeaae4c370eac39e8dcb2b724237941aaec578d80f3cf49eb"} Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.038252 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vmz8j" event={"ID":"7219ceec-2cc3-4663-bf50-11ef855ec612","Type":"ContainerStarted","Data":"2ff57f899e5e4019e8e5e1885ffa6641e464c6b7abe6c715acfe52a8a4c9b5a1"} Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.038942 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.038989 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.039816 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" event={"ID":"ae0ab229-8baf-4c5b-a452-2fdd714946ea","Type":"ContainerStarted","Data":"241fbde969d71ef2c5c267eb26d0b60189bf3f71d3394a529b4a71e42c270e44"} Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.072098 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b5847d797-hxwsr"] Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.073495 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:23 crc kubenswrapper[4946]: W1203 07:05:23.078196 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod741e35a4_2042_495b_853c_4b87a1f6d5bc.slice/crio-4b12a41902f3233c0081e00989ad2c3ce83a610421e10b95c646b41d9b196e29 WatchSource:0}: Error finding container 4b12a41902f3233c0081e00989ad2c3ce83a610421e10b95c646b41d9b196e29: Status 404 returned error can't find the container with id 4b12a41902f3233c0081e00989ad2c3ce83a610421e10b95c646b41d9b196e29 Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.088309 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:23 crc kubenswrapper[4946]: I1203 07:05:23.245552 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9"] Dec 03 07:05:23 crc kubenswrapper[4946]: W1203 07:05:23.250760 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf636b80e_63df_4797_9720_bc7daa273204.slice/crio-2fdd44b7fbd779354baf9eaf1969fe9d9baa9697031cebbbb83e4efd39349cd2 WatchSource:0}: Error finding container 2fdd44b7fbd779354baf9eaf1969fe9d9baa9697031cebbbb83e4efd39349cd2: Status 404 returned error can't find the container with id 2fdd44b7fbd779354baf9eaf1969fe9d9baa9697031cebbbb83e4efd39349cd2 Dec 03 07:05:24 crc kubenswrapper[4946]: I1203 07:05:24.049882 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b5847d797-hxwsr" event={"ID":"741e35a4-2042-495b-853c-4b87a1f6d5bc","Type":"ContainerStarted","Data":"ab72cba3b2674483c320aeeb34e3933482a18b68f0d3e10db47c579eb8ad4b84"} Dec 03 07:05:24 crc kubenswrapper[4946]: I1203 07:05:24.050246 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b5847d797-hxwsr" event={"ID":"741e35a4-2042-495b-853c-4b87a1f6d5bc","Type":"ContainerStarted","Data":"4b12a41902f3233c0081e00989ad2c3ce83a610421e10b95c646b41d9b196e29"} Dec 03 07:05:24 crc kubenswrapper[4946]: I1203 07:05:24.052042 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" event={"ID":"f636b80e-63df-4797-9720-bc7daa273204","Type":"ContainerStarted","Data":"2fdd44b7fbd779354baf9eaf1969fe9d9baa9697031cebbbb83e4efd39349cd2"} Dec 03 07:05:24 crc kubenswrapper[4946]: I1203 07:05:24.066873 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6b5847d797-hxwsr" podStartSLOduration=2.066852734 podStartE2EDuration="2.066852734s" podCreationTimestamp="2025-12-03 07:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:05:24.065225321 +0000 UTC m=+916.861915440" watchObservedRunningTime="2025-12-03 07:05:24.066852734 +0000 UTC m=+916.863542863" Dec 03 07:05:24 crc kubenswrapper[4946]: I1203 07:05:24.699611 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k5vsn"] Dec 03 07:05:25 crc kubenswrapper[4946]: I1203 07:05:25.057139 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k5vsn" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerName="registry-server" containerID="cri-o://617c377fb5733c3f58d40869c0146a64f08b8fd148cee24ee0a0b629a1871954" gracePeriod=2 Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.070063 4946 generic.go:334] "Generic (PLEG): container finished" podID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerID="617c377fb5733c3f58d40869c0146a64f08b8fd148cee24ee0a0b629a1871954" exitCode=0 Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.070153 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5vsn" event={"ID":"cd72718c-9315-41b5-8f52-edfa443eb98f","Type":"ContainerDied","Data":"617c377fb5733c3f58d40869c0146a64f08b8fd148cee24ee0a0b629a1871954"} Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.071679 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vmz8j" event={"ID":"7219ceec-2cc3-4663-bf50-11ef855ec612","Type":"ContainerStarted","Data":"cf34559276fc235673cb6934ac26f2fc61d7b8c53f83c5e9e715e76492dc72f1"} Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.071778 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.073438 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" event={"ID":"ae0ab229-8baf-4c5b-a452-2fdd714946ea","Type":"ContainerStarted","Data":"f0b337863ff9fe4660f79c4b83a24f5e97a88c53931d6e3350abe84741a6a9e8"} Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.074927 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" event={"ID":"f636b80e-63df-4797-9720-bc7daa273204","Type":"ContainerStarted","Data":"415d8194de76898f0d7521caba22918ec60d5c97982a7c2cc31f51eb85b4250e"} Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.075080 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.076222 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" event={"ID":"1e225c74-69b1-4094-9310-7e790d712264","Type":"ContainerStarted","Data":"d4570e100d10f9331c4dc4c0cddcb5d246c676d244af89bf0ed95dece2010517"} Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.092888 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vmz8j" podStartSLOduration=0.967093276 podStartE2EDuration="5.092870157s" podCreationTimestamp="2025-12-03 07:05:22 +0000 UTC" firstStartedPulling="2025-12-03 07:05:22.521646212 +0000 UTC m=+915.318336321" lastFinishedPulling="2025-12-03 07:05:26.647423043 +0000 UTC m=+919.444113202" observedRunningTime="2025-12-03 07:05:27.091139281 +0000 UTC m=+919.887829410" watchObservedRunningTime="2025-12-03 07:05:27.092870157 +0000 UTC m=+919.889560286" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.139200 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5jbpq" podStartSLOduration=1.288159046 podStartE2EDuration="5.139170684s" podCreationTimestamp="2025-12-03 07:05:22 +0000 UTC" firstStartedPulling="2025-12-03 07:05:22.790685202 +0000 UTC m=+915.587375321" lastFinishedPulling="2025-12-03 07:05:26.64169684 +0000 UTC m=+919.438386959" observedRunningTime="2025-12-03 07:05:27.134569551 +0000 UTC m=+919.931259690" watchObservedRunningTime="2025-12-03 07:05:27.139170684 +0000 UTC m=+919.935860833" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.141719 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" podStartSLOduration=1.745529088 podStartE2EDuration="5.141698052s" podCreationTimestamp="2025-12-03 07:05:22 +0000 UTC" firstStartedPulling="2025-12-03 07:05:23.254088385 +0000 UTC m=+916.050778494" lastFinishedPulling="2025-12-03 07:05:26.650257319 +0000 UTC m=+919.446947458" observedRunningTime="2025-12-03 07:05:27.109140802 +0000 UTC m=+919.905830921" watchObservedRunningTime="2025-12-03 07:05:27.141698052 +0000 UTC m=+919.938388201" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.425851 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.552910 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-catalog-content\") pod \"cd72718c-9315-41b5-8f52-edfa443eb98f\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.553029 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-utilities\") pod \"cd72718c-9315-41b5-8f52-edfa443eb98f\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.553075 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-479vl\" (UniqueName: \"kubernetes.io/projected/cd72718c-9315-41b5-8f52-edfa443eb98f-kube-api-access-479vl\") pod \"cd72718c-9315-41b5-8f52-edfa443eb98f\" (UID: \"cd72718c-9315-41b5-8f52-edfa443eb98f\") " Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.556054 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-utilities" (OuterVolumeSpecName: "utilities") pod "cd72718c-9315-41b5-8f52-edfa443eb98f" (UID: "cd72718c-9315-41b5-8f52-edfa443eb98f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.561621 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd72718c-9315-41b5-8f52-edfa443eb98f-kube-api-access-479vl" (OuterVolumeSpecName: "kube-api-access-479vl") pod "cd72718c-9315-41b5-8f52-edfa443eb98f" (UID: "cd72718c-9315-41b5-8f52-edfa443eb98f"). InnerVolumeSpecName "kube-api-access-479vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.655333 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.655379 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-479vl\" (UniqueName: \"kubernetes.io/projected/cd72718c-9315-41b5-8f52-edfa443eb98f-kube-api-access-479vl\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.723984 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd72718c-9315-41b5-8f52-edfa443eb98f" (UID: "cd72718c-9315-41b5-8f52-edfa443eb98f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:27 crc kubenswrapper[4946]: I1203 07:05:27.757070 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd72718c-9315-41b5-8f52-edfa443eb98f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:28 crc kubenswrapper[4946]: I1203 07:05:28.087513 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k5vsn" Dec 03 07:05:28 crc kubenswrapper[4946]: I1203 07:05:28.087899 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5vsn" event={"ID":"cd72718c-9315-41b5-8f52-edfa443eb98f","Type":"ContainerDied","Data":"84ab2691e2c22278deb1519bfe939a1376dc020ebe655ba862d57de6719ee263"} Dec 03 07:05:28 crc kubenswrapper[4946]: I1203 07:05:28.087987 4946 scope.go:117] "RemoveContainer" containerID="617c377fb5733c3f58d40869c0146a64f08b8fd148cee24ee0a0b629a1871954" Dec 03 07:05:28 crc kubenswrapper[4946]: I1203 07:05:28.133899 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k5vsn"] Dec 03 07:05:28 crc kubenswrapper[4946]: I1203 07:05:28.134472 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k5vsn"] Dec 03 07:05:28 crc kubenswrapper[4946]: I1203 07:05:28.326010 4946 scope.go:117] "RemoveContainer" containerID="19685f4c4ddb5cdca9aff7c556c3c83c34d8b6fb47a59766e3cd35713df8578d" Dec 03 07:05:28 crc kubenswrapper[4946]: I1203 07:05:28.347822 4946 scope.go:117] "RemoveContainer" containerID="b684e8bd47f03c7bcbbd79d073124e78f7947f20c6a8313c5769f8bbc94d78d8" Dec 03 07:05:29 crc kubenswrapper[4946]: I1203 07:05:29.600808 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" path="/var/lib/kubelet/pods/cd72718c-9315-41b5-8f52-edfa443eb98f/volumes" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.101793 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" event={"ID":"1e225c74-69b1-4094-9310-7e790d712264","Type":"ContainerStarted","Data":"b08a285a0bc6823d2ccdeb2348fd69c9303a0cd265d3e18d04d1d7742a31534c"} Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.120969 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-smzvj" podStartSLOduration=1.079559201 podStartE2EDuration="8.120941224s" podCreationTimestamp="2025-12-03 07:05:22 +0000 UTC" firstStartedPulling="2025-12-03 07:05:22.736472623 +0000 UTC m=+915.533162732" lastFinishedPulling="2025-12-03 07:05:29.777854646 +0000 UTC m=+922.574544755" observedRunningTime="2025-12-03 07:05:30.119814694 +0000 UTC m=+922.916504843" watchObservedRunningTime="2025-12-03 07:05:30.120941224 +0000 UTC m=+922.917631373" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.515861 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mwgbc"] Dec 03 07:05:30 crc kubenswrapper[4946]: E1203 07:05:30.516179 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerName="extract-content" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.516204 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerName="extract-content" Dec 03 07:05:30 crc kubenswrapper[4946]: E1203 07:05:30.516227 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerName="registry-server" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.516243 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerName="registry-server" Dec 03 07:05:30 crc kubenswrapper[4946]: E1203 07:05:30.516602 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerName="extract-utilities" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.516640 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerName="extract-utilities" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.516922 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd72718c-9315-41b5-8f52-edfa443eb98f" containerName="registry-server" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.518635 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.534610 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mwgbc"] Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.601712 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqlfw\" (UniqueName: \"kubernetes.io/projected/65659f87-a37c-424e-b917-f7ae2625a599-kube-api-access-sqlfw\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.601775 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-catalog-content\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.601851 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-utilities\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.702728 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqlfw\" (UniqueName: \"kubernetes.io/projected/65659f87-a37c-424e-b917-f7ae2625a599-kube-api-access-sqlfw\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.702795 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-catalog-content\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.702872 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-utilities\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.703348 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-utilities\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.703402 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-catalog-content\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.735311 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqlfw\" (UniqueName: \"kubernetes.io/projected/65659f87-a37c-424e-b917-f7ae2625a599-kube-api-access-sqlfw\") pod \"community-operators-mwgbc\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:30 crc kubenswrapper[4946]: I1203 07:05:30.843221 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:31 crc kubenswrapper[4946]: I1203 07:05:31.097259 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mwgbc"] Dec 03 07:05:31 crc kubenswrapper[4946]: W1203 07:05:31.103055 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65659f87_a37c_424e_b917_f7ae2625a599.slice/crio-8835c7fc30c4b4e0a26b98b9cc8420b98c3370592fef4898cf218cd26e82da3a WatchSource:0}: Error finding container 8835c7fc30c4b4e0a26b98b9cc8420b98c3370592fef4898cf218cd26e82da3a: Status 404 returned error can't find the container with id 8835c7fc30c4b4e0a26b98b9cc8420b98c3370592fef4898cf218cd26e82da3a Dec 03 07:05:32 crc kubenswrapper[4946]: I1203 07:05:32.134591 4946 generic.go:334] "Generic (PLEG): container finished" podID="65659f87-a37c-424e-b917-f7ae2625a599" containerID="19058becb7f184905a14a2d9848e71ec8e9135d7f0d49efb605b271890f64f8d" exitCode=0 Dec 03 07:05:32 crc kubenswrapper[4946]: I1203 07:05:32.137498 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgbc" event={"ID":"65659f87-a37c-424e-b917-f7ae2625a599","Type":"ContainerDied","Data":"19058becb7f184905a14a2d9848e71ec8e9135d7f0d49efb605b271890f64f8d"} Dec 03 07:05:32 crc kubenswrapper[4946]: I1203 07:05:32.137835 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgbc" event={"ID":"65659f87-a37c-424e-b917-f7ae2625a599","Type":"ContainerStarted","Data":"8835c7fc30c4b4e0a26b98b9cc8420b98c3370592fef4898cf218cd26e82da3a"} Dec 03 07:05:32 crc kubenswrapper[4946]: I1203 07:05:32.541116 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vmz8j" Dec 03 07:05:32 crc kubenswrapper[4946]: I1203 07:05:32.881130 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:32 crc kubenswrapper[4946]: I1203 07:05:32.881287 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:32 crc kubenswrapper[4946]: I1203 07:05:32.885208 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:33 crc kubenswrapper[4946]: I1203 07:05:33.144385 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgbc" event={"ID":"65659f87-a37c-424e-b917-f7ae2625a599","Type":"ContainerStarted","Data":"efb7cdb685eda6a4b35a1dacf1b4b4fb1216e8f3708f2099956e9a15ab56f247"} Dec 03 07:05:33 crc kubenswrapper[4946]: I1203 07:05:33.147867 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6b5847d797-hxwsr" Dec 03 07:05:33 crc kubenswrapper[4946]: I1203 07:05:33.214576 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zkld2"] Dec 03 07:05:34 crc kubenswrapper[4946]: I1203 07:05:34.154195 4946 generic.go:334] "Generic (PLEG): container finished" podID="65659f87-a37c-424e-b917-f7ae2625a599" containerID="efb7cdb685eda6a4b35a1dacf1b4b4fb1216e8f3708f2099956e9a15ab56f247" exitCode=0 Dec 03 07:05:34 crc kubenswrapper[4946]: I1203 07:05:34.154324 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgbc" event={"ID":"65659f87-a37c-424e-b917-f7ae2625a599","Type":"ContainerDied","Data":"efb7cdb685eda6a4b35a1dacf1b4b4fb1216e8f3708f2099956e9a15ab56f247"} Dec 03 07:05:35 crc kubenswrapper[4946]: I1203 07:05:35.162551 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgbc" event={"ID":"65659f87-a37c-424e-b917-f7ae2625a599","Type":"ContainerStarted","Data":"173861c655ebcb08b96f5c5c0256e731419bdafcbf7f9928775dbf35f0165b83"} Dec 03 07:05:35 crc kubenswrapper[4946]: I1203 07:05:35.182635 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mwgbc" podStartSLOduration=2.732565292 podStartE2EDuration="5.182609894s" podCreationTimestamp="2025-12-03 07:05:30 +0000 UTC" firstStartedPulling="2025-12-03 07:05:32.143578503 +0000 UTC m=+924.940268642" lastFinishedPulling="2025-12-03 07:05:34.593623135 +0000 UTC m=+927.390313244" observedRunningTime="2025-12-03 07:05:35.179173732 +0000 UTC m=+927.975863841" watchObservedRunningTime="2025-12-03 07:05:35.182609894 +0000 UTC m=+927.979300013" Dec 03 07:05:40 crc kubenswrapper[4946]: I1203 07:05:40.844988 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:40 crc kubenswrapper[4946]: I1203 07:05:40.845848 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:40 crc kubenswrapper[4946]: I1203 07:05:40.921919 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:41 crc kubenswrapper[4946]: I1203 07:05:41.257290 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:41 crc kubenswrapper[4946]: I1203 07:05:41.293538 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mwgbc"] Dec 03 07:05:43 crc kubenswrapper[4946]: I1203 07:05:43.083419 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jx9l9" Dec 03 07:05:43 crc kubenswrapper[4946]: I1203 07:05:43.226073 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mwgbc" podUID="65659f87-a37c-424e-b917-f7ae2625a599" containerName="registry-server" containerID="cri-o://173861c655ebcb08b96f5c5c0256e731419bdafcbf7f9928775dbf35f0165b83" gracePeriod=2 Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.234237 4946 generic.go:334] "Generic (PLEG): container finished" podID="65659f87-a37c-424e-b917-f7ae2625a599" containerID="173861c655ebcb08b96f5c5c0256e731419bdafcbf7f9928775dbf35f0165b83" exitCode=0 Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.234276 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgbc" event={"ID":"65659f87-a37c-424e-b917-f7ae2625a599","Type":"ContainerDied","Data":"173861c655ebcb08b96f5c5c0256e731419bdafcbf7f9928775dbf35f0165b83"} Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.584268 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4nvfh"] Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.586712 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.589555 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4nvfh"] Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.695942 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw8dv\" (UniqueName: \"kubernetes.io/projected/c4d42595-1772-4794-94ba-2098c5ab8a46-kube-api-access-zw8dv\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.695987 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-utilities\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.696079 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-catalog-content\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.700998 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.797502 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw8dv\" (UniqueName: \"kubernetes.io/projected/c4d42595-1772-4794-94ba-2098c5ab8a46-kube-api-access-zw8dv\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.797919 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-utilities\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.798397 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-utilities\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.798581 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-catalog-content\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.799204 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-catalog-content\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.817801 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw8dv\" (UniqueName: \"kubernetes.io/projected/c4d42595-1772-4794-94ba-2098c5ab8a46-kube-api-access-zw8dv\") pod \"certified-operators-4nvfh\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.899198 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-utilities\") pod \"65659f87-a37c-424e-b917-f7ae2625a599\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.899255 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-catalog-content\") pod \"65659f87-a37c-424e-b917-f7ae2625a599\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.899328 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqlfw\" (UniqueName: \"kubernetes.io/projected/65659f87-a37c-424e-b917-f7ae2625a599-kube-api-access-sqlfw\") pod \"65659f87-a37c-424e-b917-f7ae2625a599\" (UID: \"65659f87-a37c-424e-b917-f7ae2625a599\") " Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.900097 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-utilities" (OuterVolumeSpecName: "utilities") pod "65659f87-a37c-424e-b917-f7ae2625a599" (UID: "65659f87-a37c-424e-b917-f7ae2625a599"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.907497 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.921873 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65659f87-a37c-424e-b917-f7ae2625a599-kube-api-access-sqlfw" (OuterVolumeSpecName: "kube-api-access-sqlfw") pod "65659f87-a37c-424e-b917-f7ae2625a599" (UID: "65659f87-a37c-424e-b917-f7ae2625a599"). InnerVolumeSpecName "kube-api-access-sqlfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:05:44 crc kubenswrapper[4946]: I1203 07:05:44.973088 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65659f87-a37c-424e-b917-f7ae2625a599" (UID: "65659f87-a37c-424e-b917-f7ae2625a599"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.005698 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqlfw\" (UniqueName: \"kubernetes.io/projected/65659f87-a37c-424e-b917-f7ae2625a599-kube-api-access-sqlfw\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.005755 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.005766 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65659f87-a37c-424e-b917-f7ae2625a599-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.233050 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4nvfh"] Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.242536 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgbc" event={"ID":"65659f87-a37c-424e-b917-f7ae2625a599","Type":"ContainerDied","Data":"8835c7fc30c4b4e0a26b98b9cc8420b98c3370592fef4898cf218cd26e82da3a"} Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.242587 4946 scope.go:117] "RemoveContainer" containerID="173861c655ebcb08b96f5c5c0256e731419bdafcbf7f9928775dbf35f0165b83" Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.242634 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwgbc" Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.262001 4946 scope.go:117] "RemoveContainer" containerID="efb7cdb685eda6a4b35a1dacf1b4b4fb1216e8f3708f2099956e9a15ab56f247" Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.274280 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mwgbc"] Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.277346 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mwgbc"] Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.292604 4946 scope.go:117] "RemoveContainer" containerID="19058becb7f184905a14a2d9848e71ec8e9135d7f0d49efb605b271890f64f8d" Dec 03 07:05:45 crc kubenswrapper[4946]: I1203 07:05:45.603470 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65659f87-a37c-424e-b917-f7ae2625a599" path="/var/lib/kubelet/pods/65659f87-a37c-424e-b917-f7ae2625a599/volumes" Dec 03 07:05:46 crc kubenswrapper[4946]: I1203 07:05:46.249861 4946 generic.go:334] "Generic (PLEG): container finished" podID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerID="233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe" exitCode=0 Dec 03 07:05:46 crc kubenswrapper[4946]: I1203 07:05:46.249949 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nvfh" event={"ID":"c4d42595-1772-4794-94ba-2098c5ab8a46","Type":"ContainerDied","Data":"233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe"} Dec 03 07:05:46 crc kubenswrapper[4946]: I1203 07:05:46.249976 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nvfh" event={"ID":"c4d42595-1772-4794-94ba-2098c5ab8a46","Type":"ContainerStarted","Data":"3933a2692f4e2c2706dc2ed7a9fe7a5ebb5c55a36654f5d467714d379b8bea78"} Dec 03 07:05:47 crc kubenswrapper[4946]: I1203 07:05:47.258666 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nvfh" event={"ID":"c4d42595-1772-4794-94ba-2098c5ab8a46","Type":"ContainerStarted","Data":"09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98"} Dec 03 07:05:48 crc kubenswrapper[4946]: I1203 07:05:48.268818 4946 generic.go:334] "Generic (PLEG): container finished" podID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerID="09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98" exitCode=0 Dec 03 07:05:48 crc kubenswrapper[4946]: I1203 07:05:48.268874 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nvfh" event={"ID":"c4d42595-1772-4794-94ba-2098c5ab8a46","Type":"ContainerDied","Data":"09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98"} Dec 03 07:05:49 crc kubenswrapper[4946]: I1203 07:05:49.278645 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nvfh" event={"ID":"c4d42595-1772-4794-94ba-2098c5ab8a46","Type":"ContainerStarted","Data":"0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0"} Dec 03 07:05:49 crc kubenswrapper[4946]: I1203 07:05:49.328880 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4nvfh" podStartSLOduration=2.79140145 podStartE2EDuration="5.328859056s" podCreationTimestamp="2025-12-03 07:05:44 +0000 UTC" firstStartedPulling="2025-12-03 07:05:46.252345535 +0000 UTC m=+939.049035684" lastFinishedPulling="2025-12-03 07:05:48.789803171 +0000 UTC m=+941.586493290" observedRunningTime="2025-12-03 07:05:49.327696495 +0000 UTC m=+942.124386644" watchObservedRunningTime="2025-12-03 07:05:49.328859056 +0000 UTC m=+942.125549165" Dec 03 07:05:53 crc kubenswrapper[4946]: I1203 07:05:53.039818 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:05:53 crc kubenswrapper[4946]: I1203 07:05:53.040359 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:05:54 crc kubenswrapper[4946]: I1203 07:05:54.908806 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:54 crc kubenswrapper[4946]: I1203 07:05:54.909556 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:54 crc kubenswrapper[4946]: I1203 07:05:54.976986 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:55 crc kubenswrapper[4946]: I1203 07:05:55.373438 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:55 crc kubenswrapper[4946]: I1203 07:05:55.411535 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4nvfh"] Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.328235 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4nvfh" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerName="registry-server" containerID="cri-o://0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0" gracePeriod=2 Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.641349 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq"] Dec 03 07:05:57 crc kubenswrapper[4946]: E1203 07:05:57.641656 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65659f87-a37c-424e-b917-f7ae2625a599" containerName="extract-content" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.641677 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="65659f87-a37c-424e-b917-f7ae2625a599" containerName="extract-content" Dec 03 07:05:57 crc kubenswrapper[4946]: E1203 07:05:57.641694 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65659f87-a37c-424e-b917-f7ae2625a599" containerName="registry-server" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.641703 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="65659f87-a37c-424e-b917-f7ae2625a599" containerName="registry-server" Dec 03 07:05:57 crc kubenswrapper[4946]: E1203 07:05:57.641718 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65659f87-a37c-424e-b917-f7ae2625a599" containerName="extract-utilities" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.641726 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="65659f87-a37c-424e-b917-f7ae2625a599" containerName="extract-utilities" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.641869 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="65659f87-a37c-424e-b917-f7ae2625a599" containerName="registry-server" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.642604 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.644606 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.667232 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq"] Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.687564 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7sdg\" (UniqueName: \"kubernetes.io/projected/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-kube-api-access-v7sdg\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.687628 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.687709 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.696158 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.788429 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-utilities\") pod \"c4d42595-1772-4794-94ba-2098c5ab8a46\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.788488 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-catalog-content\") pod \"c4d42595-1772-4794-94ba-2098c5ab8a46\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.788514 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw8dv\" (UniqueName: \"kubernetes.io/projected/c4d42595-1772-4794-94ba-2098c5ab8a46-kube-api-access-zw8dv\") pod \"c4d42595-1772-4794-94ba-2098c5ab8a46\" (UID: \"c4d42595-1772-4794-94ba-2098c5ab8a46\") " Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.788707 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7sdg\" (UniqueName: \"kubernetes.io/projected/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-kube-api-access-v7sdg\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.788733 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.788788 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.789230 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.789490 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.789754 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-utilities" (OuterVolumeSpecName: "utilities") pod "c4d42595-1772-4794-94ba-2098c5ab8a46" (UID: "c4d42595-1772-4794-94ba-2098c5ab8a46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.796170 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4d42595-1772-4794-94ba-2098c5ab8a46-kube-api-access-zw8dv" (OuterVolumeSpecName: "kube-api-access-zw8dv") pod "c4d42595-1772-4794-94ba-2098c5ab8a46" (UID: "c4d42595-1772-4794-94ba-2098c5ab8a46"). InnerVolumeSpecName "kube-api-access-zw8dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.814864 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7sdg\" (UniqueName: \"kubernetes.io/projected/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-kube-api-access-v7sdg\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.890685 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.890730 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw8dv\" (UniqueName: \"kubernetes.io/projected/c4d42595-1772-4794-94ba-2098c5ab8a46-kube-api-access-zw8dv\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:57 crc kubenswrapper[4946]: I1203 07:05:57.995076 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.038563 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4d42595-1772-4794-94ba-2098c5ab8a46" (UID: "c4d42595-1772-4794-94ba-2098c5ab8a46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.093789 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4d42595-1772-4794-94ba-2098c5ab8a46-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.252987 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-zkld2" podUID="a89bae42-673b-4eb1-87d2-ef2dd919c4d4" containerName="console" containerID="cri-o://b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf" gracePeriod=15 Dec 03 07:05:58 crc kubenswrapper[4946]: W1203 07:05:58.266140 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod503cf5b8_948c_468a_a0d2_d17c0b2e7fbf.slice/crio-2166018bfe8088a36c258fbfd47d480a3d62074c0be923de795e6241a87a62c8 WatchSource:0}: Error finding container 2166018bfe8088a36c258fbfd47d480a3d62074c0be923de795e6241a87a62c8: Status 404 returned error can't find the container with id 2166018bfe8088a36c258fbfd47d480a3d62074c0be923de795e6241a87a62c8 Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.266498 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq"] Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.337369 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" event={"ID":"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf","Type":"ContainerStarted","Data":"2166018bfe8088a36c258fbfd47d480a3d62074c0be923de795e6241a87a62c8"} Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.340408 4946 generic.go:334] "Generic (PLEG): container finished" podID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerID="0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0" exitCode=0 Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.340453 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nvfh" event={"ID":"c4d42595-1772-4794-94ba-2098c5ab8a46","Type":"ContainerDied","Data":"0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0"} Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.340827 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nvfh" event={"ID":"c4d42595-1772-4794-94ba-2098c5ab8a46","Type":"ContainerDied","Data":"3933a2692f4e2c2706dc2ed7a9fe7a5ebb5c55a36654f5d467714d379b8bea78"} Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.341003 4946 scope.go:117] "RemoveContainer" containerID="0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.340488 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nvfh" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.360396 4946 scope.go:117] "RemoveContainer" containerID="09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.380013 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4nvfh"] Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.388164 4946 scope.go:117] "RemoveContainer" containerID="233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.388307 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4nvfh"] Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.405474 4946 scope.go:117] "RemoveContainer" containerID="0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0" Dec 03 07:05:58 crc kubenswrapper[4946]: E1203 07:05:58.406032 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0\": container with ID starting with 0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0 not found: ID does not exist" containerID="0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.406138 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0"} err="failed to get container status \"0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0\": rpc error: code = NotFound desc = could not find container \"0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0\": container with ID starting with 0f9f524834bccb15bb89ae5bd5b9a09f67d943e1662e5e024a0d8c72f59089b0 not found: ID does not exist" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.406225 4946 scope.go:117] "RemoveContainer" containerID="09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98" Dec 03 07:05:58 crc kubenswrapper[4946]: E1203 07:05:58.406694 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98\": container with ID starting with 09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98 not found: ID does not exist" containerID="09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.406749 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98"} err="failed to get container status \"09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98\": rpc error: code = NotFound desc = could not find container \"09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98\": container with ID starting with 09625166e809584480e27984870be681cc3cc36ed9db8d6abfe458db5451ad98 not found: ID does not exist" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.406779 4946 scope.go:117] "RemoveContainer" containerID="233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe" Dec 03 07:05:58 crc kubenswrapper[4946]: E1203 07:05:58.407257 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe\": container with ID starting with 233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe not found: ID does not exist" containerID="233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe" Dec 03 07:05:58 crc kubenswrapper[4946]: I1203 07:05:58.407295 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe"} err="failed to get container status \"233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe\": rpc error: code = NotFound desc = could not find container \"233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe\": container with ID starting with 233fa65c4b760a38364680b02ec64059b5aa988825aa2e09795e4f225283e8fe not found: ID does not exist" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.245217 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zkld2_a89bae42-673b-4eb1-87d2-ef2dd919c4d4/console/0.log" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.245541 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkld2" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.311809 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-serving-cert\") pod \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.311886 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gssj8\" (UniqueName: \"kubernetes.io/projected/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-kube-api-access-gssj8\") pod \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.311939 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-oauth-serving-cert\") pod \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.311990 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-service-ca\") pod \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.312049 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-oauth-config\") pod \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.312087 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-trusted-ca-bundle\") pod \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.312143 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-config\") pod \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\" (UID: \"a89bae42-673b-4eb1-87d2-ef2dd919c4d4\") " Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.312975 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a89bae42-673b-4eb1-87d2-ef2dd919c4d4" (UID: "a89bae42-673b-4eb1-87d2-ef2dd919c4d4"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.313186 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-config" (OuterVolumeSpecName: "console-config") pod "a89bae42-673b-4eb1-87d2-ef2dd919c4d4" (UID: "a89bae42-673b-4eb1-87d2-ef2dd919c4d4"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.313273 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a89bae42-673b-4eb1-87d2-ef2dd919c4d4" (UID: "a89bae42-673b-4eb1-87d2-ef2dd919c4d4"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.313262 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-service-ca" (OuterVolumeSpecName: "service-ca") pod "a89bae42-673b-4eb1-87d2-ef2dd919c4d4" (UID: "a89bae42-673b-4eb1-87d2-ef2dd919c4d4"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.318719 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a89bae42-673b-4eb1-87d2-ef2dd919c4d4" (UID: "a89bae42-673b-4eb1-87d2-ef2dd919c4d4"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.319283 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a89bae42-673b-4eb1-87d2-ef2dd919c4d4" (UID: "a89bae42-673b-4eb1-87d2-ef2dd919c4d4"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.319490 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-kube-api-access-gssj8" (OuterVolumeSpecName: "kube-api-access-gssj8") pod "a89bae42-673b-4eb1-87d2-ef2dd919c4d4" (UID: "a89bae42-673b-4eb1-87d2-ef2dd919c4d4"). InnerVolumeSpecName "kube-api-access-gssj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.349138 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zkld2_a89bae42-673b-4eb1-87d2-ef2dd919c4d4/console/0.log" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.349179 4946 generic.go:334] "Generic (PLEG): container finished" podID="a89bae42-673b-4eb1-87d2-ef2dd919c4d4" containerID="b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf" exitCode=2 Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.349226 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkld2" event={"ID":"a89bae42-673b-4eb1-87d2-ef2dd919c4d4","Type":"ContainerDied","Data":"b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf"} Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.349248 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkld2" event={"ID":"a89bae42-673b-4eb1-87d2-ef2dd919c4d4","Type":"ContainerDied","Data":"bf44b5944aec0bb6730f8813277691ed0968ac0e2d0da5c4ed06ef9b5f012c53"} Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.349263 4946 scope.go:117] "RemoveContainer" containerID="b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.349275 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkld2" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.353444 4946 generic.go:334] "Generic (PLEG): container finished" podID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerID="ef6a126c33b06b723f6722ec74b23b2e760fa569ac8253bdd7bb8a8e629b2bc9" exitCode=0 Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.353491 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" event={"ID":"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf","Type":"ContainerDied","Data":"ef6a126c33b06b723f6722ec74b23b2e760fa569ac8253bdd7bb8a8e629b2bc9"} Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.379841 4946 scope.go:117] "RemoveContainer" containerID="b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf" Dec 03 07:05:59 crc kubenswrapper[4946]: E1203 07:05:59.380955 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf\": container with ID starting with b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf not found: ID does not exist" containerID="b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.381027 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf"} err="failed to get container status \"b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf\": rpc error: code = NotFound desc = could not find container \"b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf\": container with ID starting with b8bda90129d415cb0fa36cdc359148b363d31ba6faac97ee3fa723e90137fdaf not found: ID does not exist" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.395469 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zkld2"] Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.403377 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-zkld2"] Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.414671 4946 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.414701 4946 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.414759 4946 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.414778 4946 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.414827 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gssj8\" (UniqueName: \"kubernetes.io/projected/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-kube-api-access-gssj8\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.414845 4946 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.414856 4946 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a89bae42-673b-4eb1-87d2-ef2dd919c4d4-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.605178 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a89bae42-673b-4eb1-87d2-ef2dd919c4d4" path="/var/lib/kubelet/pods/a89bae42-673b-4eb1-87d2-ef2dd919c4d4/volumes" Dec 03 07:05:59 crc kubenswrapper[4946]: I1203 07:05:59.605830 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" path="/var/lib/kubelet/pods/c4d42595-1772-4794-94ba-2098c5ab8a46/volumes" Dec 03 07:06:00 crc kubenswrapper[4946]: I1203 07:06:00.364062 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" event={"ID":"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf","Type":"ContainerStarted","Data":"ede1f0d7f81f4d425568bd5db75113d1ba57c03ab5f8e368cb019f01c8e2103b"} Dec 03 07:06:01 crc kubenswrapper[4946]: I1203 07:06:01.374644 4946 generic.go:334] "Generic (PLEG): container finished" podID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerID="ede1f0d7f81f4d425568bd5db75113d1ba57c03ab5f8e368cb019f01c8e2103b" exitCode=0 Dec 03 07:06:01 crc kubenswrapper[4946]: I1203 07:06:01.374721 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" event={"ID":"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf","Type":"ContainerDied","Data":"ede1f0d7f81f4d425568bd5db75113d1ba57c03ab5f8e368cb019f01c8e2103b"} Dec 03 07:06:02 crc kubenswrapper[4946]: I1203 07:06:02.386340 4946 generic.go:334] "Generic (PLEG): container finished" podID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerID="e8208b0506c3aab31360616f54073e4c478cc12b15bbde7541dfd6bba35ed2b0" exitCode=0 Dec 03 07:06:02 crc kubenswrapper[4946]: I1203 07:06:02.386406 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" event={"ID":"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf","Type":"ContainerDied","Data":"e8208b0506c3aab31360616f54073e4c478cc12b15bbde7541dfd6bba35ed2b0"} Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.686168 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.774566 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-bundle\") pod \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.774892 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7sdg\" (UniqueName: \"kubernetes.io/projected/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-kube-api-access-v7sdg\") pod \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.774990 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-util\") pod \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\" (UID: \"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf\") " Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.776275 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-bundle" (OuterVolumeSpecName: "bundle") pod "503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" (UID: "503cf5b8-948c-468a-a0d2-d17c0b2e7fbf"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.782449 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-kube-api-access-v7sdg" (OuterVolumeSpecName: "kube-api-access-v7sdg") pod "503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" (UID: "503cf5b8-948c-468a-a0d2-d17c0b2e7fbf"). InnerVolumeSpecName "kube-api-access-v7sdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.811500 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-util" (OuterVolumeSpecName: "util") pod "503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" (UID: "503cf5b8-948c-468a-a0d2-d17c0b2e7fbf"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.876373 4946 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.876407 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7sdg\" (UniqueName: \"kubernetes.io/projected/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-kube-api-access-v7sdg\") on node \"crc\" DevicePath \"\"" Dec 03 07:06:03 crc kubenswrapper[4946]: I1203 07:06:03.876421 4946 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/503cf5b8-948c-468a-a0d2-d17c0b2e7fbf-util\") on node \"crc\" DevicePath \"\"" Dec 03 07:06:04 crc kubenswrapper[4946]: I1203 07:06:04.403454 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" event={"ID":"503cf5b8-948c-468a-a0d2-d17c0b2e7fbf","Type":"ContainerDied","Data":"2166018bfe8088a36c258fbfd47d480a3d62074c0be923de795e6241a87a62c8"} Dec 03 07:06:04 crc kubenswrapper[4946]: I1203 07:06:04.403512 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2166018bfe8088a36c258fbfd47d480a3d62074c0be923de795e6241a87a62c8" Dec 03 07:06:04 crc kubenswrapper[4946]: I1203 07:06:04.403520 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.574226 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5597d58577-85fkn"] Dec 03 07:06:14 crc kubenswrapper[4946]: E1203 07:06:14.575001 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89bae42-673b-4eb1-87d2-ef2dd919c4d4" containerName="console" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575014 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89bae42-673b-4eb1-87d2-ef2dd919c4d4" containerName="console" Dec 03 07:06:14 crc kubenswrapper[4946]: E1203 07:06:14.575020 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerName="registry-server" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575026 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerName="registry-server" Dec 03 07:06:14 crc kubenswrapper[4946]: E1203 07:06:14.575036 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerName="util" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575042 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerName="util" Dec 03 07:06:14 crc kubenswrapper[4946]: E1203 07:06:14.575053 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerName="extract" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575059 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerName="extract" Dec 03 07:06:14 crc kubenswrapper[4946]: E1203 07:06:14.575068 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerName="extract-content" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575073 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerName="extract-content" Dec 03 07:06:14 crc kubenswrapper[4946]: E1203 07:06:14.575083 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerName="pull" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575089 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerName="pull" Dec 03 07:06:14 crc kubenswrapper[4946]: E1203 07:06:14.575097 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerName="extract-utilities" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575103 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerName="extract-utilities" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575209 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4d42595-1772-4794-94ba-2098c5ab8a46" containerName="registry-server" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575220 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89bae42-673b-4eb1-87d2-ef2dd919c4d4" containerName="console" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575232 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="503cf5b8-948c-468a-a0d2-d17c0b2e7fbf" containerName="extract" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.575561 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.577373 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.577402 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.577373 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.577628 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.578544 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-t4bf5" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.589951 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5597d58577-85fkn"] Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.708965 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpk9b\" (UniqueName: \"kubernetes.io/projected/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-kube-api-access-fpk9b\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.709102 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-apiservice-cert\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.709180 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-webhook-cert\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.810086 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-apiservice-cert\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.810171 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-webhook-cert\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.810214 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpk9b\" (UniqueName: \"kubernetes.io/projected/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-kube-api-access-fpk9b\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.816538 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-apiservice-cert\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.827546 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-webhook-cert\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.831727 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpk9b\" (UniqueName: \"kubernetes.io/projected/9cefae38-3773-4b3c-b0a7-fd4c4c3fea14-kube-api-access-fpk9b\") pod \"metallb-operator-controller-manager-5597d58577-85fkn\" (UID: \"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14\") " pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.893200 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.921950 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw"] Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.922606 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.927216 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-crf8b" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.927281 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.927216 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 07:06:14 crc kubenswrapper[4946]: I1203 07:06:14.971200 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw"] Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.012503 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-apiservice-cert\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.012567 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-webhook-cert\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.012588 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkt2b\" (UniqueName: \"kubernetes.io/projected/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-kube-api-access-fkt2b\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.113574 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-webhook-cert\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.113616 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkt2b\" (UniqueName: \"kubernetes.io/projected/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-kube-api-access-fkt2b\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.113669 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-apiservice-cert\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.128129 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-webhook-cert\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.130527 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-apiservice-cert\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.135482 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkt2b\" (UniqueName: \"kubernetes.io/projected/57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d-kube-api-access-fkt2b\") pod \"metallb-operator-webhook-server-78b68b94b4-klptw\" (UID: \"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d\") " pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.178979 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5597d58577-85fkn"] Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.292154 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.488028 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw"] Dec 03 07:06:15 crc kubenswrapper[4946]: W1203 07:06:15.494145 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57c3ab00_3c9b_41d1_8c7e_9c51b8fd5b3d.slice/crio-04de6ee94ed818e6c42fd72286d3a19e744349c86abf0459cca4a7ab9caf71f0 WatchSource:0}: Error finding container 04de6ee94ed818e6c42fd72286d3a19e744349c86abf0459cca4a7ab9caf71f0: Status 404 returned error can't find the container with id 04de6ee94ed818e6c42fd72286d3a19e744349c86abf0459cca4a7ab9caf71f0 Dec 03 07:06:15 crc kubenswrapper[4946]: I1203 07:06:15.497838 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" event={"ID":"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14","Type":"ContainerStarted","Data":"ba7aca7f7122de1553921643e549035d8adb1aababf246c6aa202cdcae89d24f"} Dec 03 07:06:16 crc kubenswrapper[4946]: I1203 07:06:16.504353 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" event={"ID":"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d","Type":"ContainerStarted","Data":"04de6ee94ed818e6c42fd72286d3a19e744349c86abf0459cca4a7ab9caf71f0"} Dec 03 07:06:19 crc kubenswrapper[4946]: I1203 07:06:19.531200 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" event={"ID":"9cefae38-3773-4b3c-b0a7-fd4c4c3fea14","Type":"ContainerStarted","Data":"b59483145e1fdcb02e466d600e48d4fdffff1bdf7abebc4cfff9ef47502b2085"} Dec 03 07:06:19 crc kubenswrapper[4946]: I1203 07:06:19.531792 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:19 crc kubenswrapper[4946]: I1203 07:06:19.567462 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" podStartSLOduration=1.6679445990000001 podStartE2EDuration="5.567439564s" podCreationTimestamp="2025-12-03 07:06:14 +0000 UTC" firstStartedPulling="2025-12-03 07:06:15.246011744 +0000 UTC m=+968.042701853" lastFinishedPulling="2025-12-03 07:06:19.145506669 +0000 UTC m=+971.942196818" observedRunningTime="2025-12-03 07:06:19.563086828 +0000 UTC m=+972.359776947" watchObservedRunningTime="2025-12-03 07:06:19.567439564 +0000 UTC m=+972.364129663" Dec 03 07:06:20 crc kubenswrapper[4946]: I1203 07:06:20.537832 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" event={"ID":"57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d","Type":"ContainerStarted","Data":"1be96715c716f0efb1dd7d4d4118e60e1d1f3f1d562f974122b7d70bb80ebb49"} Dec 03 07:06:20 crc kubenswrapper[4946]: I1203 07:06:20.559893 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" podStartSLOduration=2.598957499 podStartE2EDuration="6.559871655s" podCreationTimestamp="2025-12-03 07:06:14 +0000 UTC" firstStartedPulling="2025-12-03 07:06:15.496606441 +0000 UTC m=+968.293296550" lastFinishedPulling="2025-12-03 07:06:19.457520597 +0000 UTC m=+972.254210706" observedRunningTime="2025-12-03 07:06:20.55707686 +0000 UTC m=+973.353766969" watchObservedRunningTime="2025-12-03 07:06:20.559871655 +0000 UTC m=+973.356561774" Dec 03 07:06:21 crc kubenswrapper[4946]: I1203 07:06:21.542630 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.048826 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.049188 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.049238 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.050171 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb200c920ce9a9189cb9d22399d35826a7f0ee55cd4d5b661511822e8281f026"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.050345 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://eb200c920ce9a9189cb9d22399d35826a7f0ee55cd4d5b661511822e8281f026" gracePeriod=600 Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.555113 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="eb200c920ce9a9189cb9d22399d35826a7f0ee55cd4d5b661511822e8281f026" exitCode=0 Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.555171 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"eb200c920ce9a9189cb9d22399d35826a7f0ee55cd4d5b661511822e8281f026"} Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.555593 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"b5f3af879642bc444185e03b298fe438ec767c4e61a69937761fbbcf8110a3b1"} Dec 03 07:06:23 crc kubenswrapper[4946]: I1203 07:06:23.555632 4946 scope.go:117] "RemoveContainer" containerID="5f5bf61ac3d07d18e68762b57ed405ab4ef0e6a554347d633d86406fb904be5c" Dec 03 07:06:35 crc kubenswrapper[4946]: I1203 07:06:35.298213 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-78b68b94b4-klptw" Dec 03 07:06:54 crc kubenswrapper[4946]: I1203 07:06:54.898615 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5597d58577-85fkn" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.789983 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx"] Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.790591 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.792467 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.792921 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7sj8q" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.812450 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-4h989"] Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.815087 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.815487 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx"] Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.816795 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.824431 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.903557 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-fsnxs"] Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.904523 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fsnxs" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.908477 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kmkf2" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.909325 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.910105 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.913719 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-5bh6f"] Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.916707 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.917162 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.918801 4946 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.942820 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-5bh6f"] Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978100 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8j5j\" (UniqueName: \"kubernetes.io/projected/55fde23e-25e3-4aca-88a3-878ae3117972-kube-api-access-j8j5j\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978140 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-reloader\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978166 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-metrics-certs\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978215 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-metrics\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978235 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-cert\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978257 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/55fde23e-25e3-4aca-88a3-878ae3117972-metallb-excludel2\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978298 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c6d5045-fb6d-407d-b044-f42503aa1ce1-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-kb5zx\" (UID: \"6c6d5045-fb6d-407d-b044-f42503aa1ce1\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978367 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-metrics-certs\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978410 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc5hl\" (UniqueName: \"kubernetes.io/projected/34f70191-7d9a-4c12-8a82-f2927ac34e75-kube-api-access-tc5hl\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978436 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34f70191-7d9a-4c12-8a82-f2927ac34e75-metrics-certs\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978453 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978472 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-sockets\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978491 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-startup\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978508 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpgw4\" (UniqueName: \"kubernetes.io/projected/6c6d5045-fb6d-407d-b044-f42503aa1ce1-kube-api-access-xpgw4\") pod \"frr-k8s-webhook-server-7fcb986d4-kb5zx\" (UID: \"6c6d5045-fb6d-407d-b044-f42503aa1ce1\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978534 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-conf\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:55 crc kubenswrapper[4946]: I1203 07:06:55.978548 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rxgr\" (UniqueName: \"kubernetes.io/projected/3ff6d4b8-37a3-43af-bac0-00a51011a791-kube-api-access-6rxgr\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079532 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-metrics\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079571 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-cert\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079594 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/55fde23e-25e3-4aca-88a3-878ae3117972-metallb-excludel2\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079623 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c6d5045-fb6d-407d-b044-f42503aa1ce1-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-kb5zx\" (UID: \"6c6d5045-fb6d-407d-b044-f42503aa1ce1\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079643 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-metrics-certs\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079663 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc5hl\" (UniqueName: \"kubernetes.io/projected/34f70191-7d9a-4c12-8a82-f2927ac34e75-kube-api-access-tc5hl\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079685 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34f70191-7d9a-4c12-8a82-f2927ac34e75-metrics-certs\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079698 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079715 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-sockets\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079746 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-startup\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079764 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpgw4\" (UniqueName: \"kubernetes.io/projected/6c6d5045-fb6d-407d-b044-f42503aa1ce1-kube-api-access-xpgw4\") pod \"frr-k8s-webhook-server-7fcb986d4-kb5zx\" (UID: \"6c6d5045-fb6d-407d-b044-f42503aa1ce1\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079787 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-conf\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.079802 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rxgr\" (UniqueName: \"kubernetes.io/projected/3ff6d4b8-37a3-43af-bac0-00a51011a791-kube-api-access-6rxgr\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:56 crc kubenswrapper[4946]: E1203 07:06:56.079954 4946 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 07:06:56 crc kubenswrapper[4946]: E1203 07:06:56.080016 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist podName:55fde23e-25e3-4aca-88a3-878ae3117972 nodeName:}" failed. No retries permitted until 2025-12-03 07:06:56.579996952 +0000 UTC m=+1009.376687051 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist") pod "speaker-fsnxs" (UID: "55fde23e-25e3-4aca-88a3-878ae3117972") : secret "metallb-memberlist" not found Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.080024 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-metrics\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.080150 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-sockets\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.080301 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-conf\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.080312 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8j5j\" (UniqueName: \"kubernetes.io/projected/55fde23e-25e3-4aca-88a3-878ae3117972-kube-api-access-j8j5j\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.080363 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-reloader\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.080395 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-metrics-certs\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:56 crc kubenswrapper[4946]: E1203 07:06:56.080477 4946 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 03 07:06:56 crc kubenswrapper[4946]: E1203 07:06:56.080534 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-metrics-certs podName:3ff6d4b8-37a3-43af-bac0-00a51011a791 nodeName:}" failed. No retries permitted until 2025-12-03 07:06:56.580519306 +0000 UTC m=+1009.377209415 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-metrics-certs") pod "controller-f8648f98b-5bh6f" (UID: "3ff6d4b8-37a3-43af-bac0-00a51011a791") : secret "controller-certs-secret" not found Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.080690 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/34f70191-7d9a-4c12-8a82-f2927ac34e75-reloader\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.080912 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/34f70191-7d9a-4c12-8a82-f2927ac34e75-frr-startup\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.081091 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/55fde23e-25e3-4aca-88a3-878ae3117972-metallb-excludel2\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.085340 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34f70191-7d9a-4c12-8a82-f2927ac34e75-metrics-certs\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.086232 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-cert\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.090246 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-metrics-certs\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.094640 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc5hl\" (UniqueName: \"kubernetes.io/projected/34f70191-7d9a-4c12-8a82-f2927ac34e75-kube-api-access-tc5hl\") pod \"frr-k8s-4h989\" (UID: \"34f70191-7d9a-4c12-8a82-f2927ac34e75\") " pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.095348 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c6d5045-fb6d-407d-b044-f42503aa1ce1-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-kb5zx\" (UID: \"6c6d5045-fb6d-407d-b044-f42503aa1ce1\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.097197 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpgw4\" (UniqueName: \"kubernetes.io/projected/6c6d5045-fb6d-407d-b044-f42503aa1ce1-kube-api-access-xpgw4\") pod \"frr-k8s-webhook-server-7fcb986d4-kb5zx\" (UID: \"6c6d5045-fb6d-407d-b044-f42503aa1ce1\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.101723 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rxgr\" (UniqueName: \"kubernetes.io/projected/3ff6d4b8-37a3-43af-bac0-00a51011a791-kube-api-access-6rxgr\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.105324 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8j5j\" (UniqueName: \"kubernetes.io/projected/55fde23e-25e3-4aca-88a3-878ae3117972-kube-api-access-j8j5j\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.107221 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.127686 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-4h989" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.512155 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx"] Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.585244 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.585367 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-metrics-certs\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:56 crc kubenswrapper[4946]: E1203 07:06:56.585808 4946 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 07:06:56 crc kubenswrapper[4946]: E1203 07:06:56.585876 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist podName:55fde23e-25e3-4aca-88a3-878ae3117972 nodeName:}" failed. No retries permitted until 2025-12-03 07:06:57.585858751 +0000 UTC m=+1010.382548870 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist") pod "speaker-fsnxs" (UID: "55fde23e-25e3-4aca-88a3-878ae3117972") : secret "metallb-memberlist" not found Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.593923 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3ff6d4b8-37a3-43af-bac0-00a51011a791-metrics-certs\") pod \"controller-f8648f98b-5bh6f\" (UID: \"3ff6d4b8-37a3-43af-bac0-00a51011a791\") " pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.767934 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" event={"ID":"6c6d5045-fb6d-407d-b044-f42503aa1ce1","Type":"ContainerStarted","Data":"f01e61656480ab1ab2080fd3735d08025335af67fac3c44fe99aed8a4c6c1b6e"} Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.769547 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerStarted","Data":"bbab98e5c5df61f6e2689798bb53ed3c6081722d30c4ec198d4d7d2364bd6c9d"} Dec 03 07:06:56 crc kubenswrapper[4946]: I1203 07:06:56.869237 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.074214 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-5bh6f"] Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.602681 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.616357 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/55fde23e-25e3-4aca-88a3-878ae3117972-memberlist\") pod \"speaker-fsnxs\" (UID: \"55fde23e-25e3-4aca-88a3-878ae3117972\") " pod="metallb-system/speaker-fsnxs" Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.759282 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fsnxs" Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.781156 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-5bh6f" event={"ID":"3ff6d4b8-37a3-43af-bac0-00a51011a791","Type":"ContainerStarted","Data":"fb0ee1b74046f203364c07b0ffc83cfc848d8515d7b96d203fcc8e3b4c5be780"} Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.781196 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-5bh6f" event={"ID":"3ff6d4b8-37a3-43af-bac0-00a51011a791","Type":"ContainerStarted","Data":"c41d381174312766843becc9d346070c0e5d000f1c06d77892d6bedc41c97edc"} Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.781206 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-5bh6f" event={"ID":"3ff6d4b8-37a3-43af-bac0-00a51011a791","Type":"ContainerStarted","Data":"da1913b84fd2e6fbbd827180fb9ed2c5e10752d0baabc8ecfddb8b51a5101c6a"} Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.781860 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:06:57 crc kubenswrapper[4946]: I1203 07:06:57.796923 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-5bh6f" podStartSLOduration=2.796906947 podStartE2EDuration="2.796906947s" podCreationTimestamp="2025-12-03 07:06:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:06:57.794995126 +0000 UTC m=+1010.591685235" watchObservedRunningTime="2025-12-03 07:06:57.796906947 +0000 UTC m=+1010.593597056" Dec 03 07:06:58 crc kubenswrapper[4946]: I1203 07:06:58.798155 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fsnxs" event={"ID":"55fde23e-25e3-4aca-88a3-878ae3117972","Type":"ContainerStarted","Data":"02317e00a3a0b647f046f3123846cefc16859c274a5a509ecefdfa574c604393"} Dec 03 07:06:58 crc kubenswrapper[4946]: I1203 07:06:58.798546 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fsnxs" event={"ID":"55fde23e-25e3-4aca-88a3-878ae3117972","Type":"ContainerStarted","Data":"da803bc741b0e028a257c0e491a91d9f34b857d7165df27d230244f84e843ef6"} Dec 03 07:06:58 crc kubenswrapper[4946]: I1203 07:06:58.798566 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fsnxs" event={"ID":"55fde23e-25e3-4aca-88a3-878ae3117972","Type":"ContainerStarted","Data":"03bf0f9e1c543317df10b5c0b7074872aa66a357d7ba0f13a3e001d1680dd06d"} Dec 03 07:06:58 crc kubenswrapper[4946]: I1203 07:06:58.798791 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-fsnxs" Dec 03 07:06:58 crc kubenswrapper[4946]: I1203 07:06:58.828395 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-fsnxs" podStartSLOduration=3.828377687 podStartE2EDuration="3.828377687s" podCreationTimestamp="2025-12-03 07:06:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:06:58.823198197 +0000 UTC m=+1011.619888306" watchObservedRunningTime="2025-12-03 07:06:58.828377687 +0000 UTC m=+1011.625067806" Dec 03 07:07:03 crc kubenswrapper[4946]: I1203 07:07:03.831644 4946 generic.go:334] "Generic (PLEG): container finished" podID="34f70191-7d9a-4c12-8a82-f2927ac34e75" containerID="ffad31ab76dfadbeb197fd46f420bfa7f0f75686cf1dc41d871158f04feaf139" exitCode=0 Dec 03 07:07:03 crc kubenswrapper[4946]: I1203 07:07:03.831721 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerDied","Data":"ffad31ab76dfadbeb197fd46f420bfa7f0f75686cf1dc41d871158f04feaf139"} Dec 03 07:07:03 crc kubenswrapper[4946]: I1203 07:07:03.834381 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" event={"ID":"6c6d5045-fb6d-407d-b044-f42503aa1ce1","Type":"ContainerStarted","Data":"30fbec7effe8351230fb5078b53323ec3e6e9b74dbfe0dffe503d7c518a3c693"} Dec 03 07:07:03 crc kubenswrapper[4946]: I1203 07:07:03.834928 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:07:03 crc kubenswrapper[4946]: I1203 07:07:03.906352 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" podStartSLOduration=2.579100602 podStartE2EDuration="8.906317771s" podCreationTimestamp="2025-12-03 07:06:55 +0000 UTC" firstStartedPulling="2025-12-03 07:06:56.519140275 +0000 UTC m=+1009.315830404" lastFinishedPulling="2025-12-03 07:07:02.846357464 +0000 UTC m=+1015.643047573" observedRunningTime="2025-12-03 07:07:03.891703728 +0000 UTC m=+1016.688393867" watchObservedRunningTime="2025-12-03 07:07:03.906317771 +0000 UTC m=+1016.703007930" Dec 03 07:07:04 crc kubenswrapper[4946]: I1203 07:07:04.842263 4946 generic.go:334] "Generic (PLEG): container finished" podID="34f70191-7d9a-4c12-8a82-f2927ac34e75" containerID="6acf3a193ac72bf30d456a0b2db31ff75d64d1d2fbea93133fbc5282500e179b" exitCode=0 Dec 03 07:07:04 crc kubenswrapper[4946]: I1203 07:07:04.842304 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerDied","Data":"6acf3a193ac72bf30d456a0b2db31ff75d64d1d2fbea93133fbc5282500e179b"} Dec 03 07:07:05 crc kubenswrapper[4946]: I1203 07:07:05.853828 4946 generic.go:334] "Generic (PLEG): container finished" podID="34f70191-7d9a-4c12-8a82-f2927ac34e75" containerID="070a150e1535b6231b4af4883e2bae52a5709d305dec91ae365d634ce87a44a9" exitCode=0 Dec 03 07:07:05 crc kubenswrapper[4946]: I1203 07:07:05.853949 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerDied","Data":"070a150e1535b6231b4af4883e2bae52a5709d305dec91ae365d634ce87a44a9"} Dec 03 07:07:06 crc kubenswrapper[4946]: I1203 07:07:06.862077 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerStarted","Data":"2391c37908913353aeddac7d5da92a70aba400372dfe9061e7dc3edbefe1e931"} Dec 03 07:07:06 crc kubenswrapper[4946]: I1203 07:07:06.862994 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerStarted","Data":"c0524006d48a662abfe1b5669ff2b7547e0e95ae0d5bc72ef5e3d0152b6db027"} Dec 03 07:07:07 crc kubenswrapper[4946]: I1203 07:07:07.874496 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerStarted","Data":"248bf9b1dced8dec0067f7a0db77cd4c5083c961951acb8250ab6107b668a346"} Dec 03 07:07:07 crc kubenswrapper[4946]: I1203 07:07:07.875159 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerStarted","Data":"1af7c3d1d9c9e6347de992bdb76ddbe717e9bda4b482246de4ee45261ea8766b"} Dec 03 07:07:07 crc kubenswrapper[4946]: I1203 07:07:07.875177 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerStarted","Data":"bc596c71647b829083630a4012a051f5c2df588344c4d0e09016cd9f295af520"} Dec 03 07:07:08 crc kubenswrapper[4946]: I1203 07:07:08.892511 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4h989" event={"ID":"34f70191-7d9a-4c12-8a82-f2927ac34e75","Type":"ContainerStarted","Data":"6c58967d0adabf088d35841f45d65dd202dd81c336531613108960217303aa3d"} Dec 03 07:07:08 crc kubenswrapper[4946]: I1203 07:07:08.894025 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-4h989" Dec 03 07:07:08 crc kubenswrapper[4946]: I1203 07:07:08.928569 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-4h989" podStartSLOduration=7.434760421 podStartE2EDuration="13.928541004s" podCreationTimestamp="2025-12-03 07:06:55 +0000 UTC" firstStartedPulling="2025-12-03 07:06:56.314818484 +0000 UTC m=+1009.111508593" lastFinishedPulling="2025-12-03 07:07:02.808599057 +0000 UTC m=+1015.605289176" observedRunningTime="2025-12-03 07:07:08.922776769 +0000 UTC m=+1021.719466928" watchObservedRunningTime="2025-12-03 07:07:08.928541004 +0000 UTC m=+1021.725231153" Dec 03 07:07:11 crc kubenswrapper[4946]: I1203 07:07:11.128958 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-4h989" Dec 03 07:07:11 crc kubenswrapper[4946]: I1203 07:07:11.169391 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-4h989" Dec 03 07:07:16 crc kubenswrapper[4946]: I1203 07:07:16.131214 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kb5zx" Dec 03 07:07:16 crc kubenswrapper[4946]: I1203 07:07:16.132889 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-4h989" Dec 03 07:07:16 crc kubenswrapper[4946]: I1203 07:07:16.875126 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-5bh6f" Dec 03 07:07:17 crc kubenswrapper[4946]: I1203 07:07:17.763218 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-fsnxs" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.379212 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj"] Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.381322 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.384287 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.394453 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj"] Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.518828 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.518921 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqth2\" (UniqueName: \"kubernetes.io/projected/268ff739-d59c-4d7c-a413-bcb696ef8c86-kube-api-access-cqth2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.518960 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.620283 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.620451 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqth2\" (UniqueName: \"kubernetes.io/projected/268ff739-d59c-4d7c-a413-bcb696ef8c86-kube-api-access-cqth2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.620534 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.620809 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.621109 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:19 crc kubenswrapper[4946]: I1203 07:07:19.877408 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqth2\" (UniqueName: \"kubernetes.io/projected/268ff739-d59c-4d7c-a413-bcb696ef8c86-kube-api-access-cqth2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:20 crc kubenswrapper[4946]: I1203 07:07:20.016449 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:20 crc kubenswrapper[4946]: I1203 07:07:20.473428 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj"] Dec 03 07:07:20 crc kubenswrapper[4946]: W1203 07:07:20.483464 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod268ff739_d59c_4d7c_a413_bcb696ef8c86.slice/crio-53a4686f428de83e7805dbecbfe5a89a0c91e762333a6828eeba07ca32ea9e84 WatchSource:0}: Error finding container 53a4686f428de83e7805dbecbfe5a89a0c91e762333a6828eeba07ca32ea9e84: Status 404 returned error can't find the container with id 53a4686f428de83e7805dbecbfe5a89a0c91e762333a6828eeba07ca32ea9e84 Dec 03 07:07:20 crc kubenswrapper[4946]: I1203 07:07:20.974075 4946 generic.go:334] "Generic (PLEG): container finished" podID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerID="6ed56f095aa445155552d17f1e8428877b3f2bc88f5512118bce13acdc6c5c0a" exitCode=0 Dec 03 07:07:20 crc kubenswrapper[4946]: I1203 07:07:20.974212 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" event={"ID":"268ff739-d59c-4d7c-a413-bcb696ef8c86","Type":"ContainerDied","Data":"6ed56f095aa445155552d17f1e8428877b3f2bc88f5512118bce13acdc6c5c0a"} Dec 03 07:07:20 crc kubenswrapper[4946]: I1203 07:07:20.974480 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" event={"ID":"268ff739-d59c-4d7c-a413-bcb696ef8c86","Type":"ContainerStarted","Data":"53a4686f428de83e7805dbecbfe5a89a0c91e762333a6828eeba07ca32ea9e84"} Dec 03 07:07:23 crc kubenswrapper[4946]: I1203 07:07:23.995427 4946 generic.go:334] "Generic (PLEG): container finished" podID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerID="25ddd75ed4cf36a13113c91366e3d05d1e4f985574da60319bef7c7e2acae02f" exitCode=0 Dec 03 07:07:23 crc kubenswrapper[4946]: I1203 07:07:23.995486 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" event={"ID":"268ff739-d59c-4d7c-a413-bcb696ef8c86","Type":"ContainerDied","Data":"25ddd75ed4cf36a13113c91366e3d05d1e4f985574da60319bef7c7e2acae02f"} Dec 03 07:07:25 crc kubenswrapper[4946]: I1203 07:07:25.006050 4946 generic.go:334] "Generic (PLEG): container finished" podID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerID="6c041eb6552696acec4cdce751ffba6f4fadcbc11c748d4d9e94cb05149a48d4" exitCode=0 Dec 03 07:07:25 crc kubenswrapper[4946]: I1203 07:07:25.006113 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" event={"ID":"268ff739-d59c-4d7c-a413-bcb696ef8c86","Type":"ContainerDied","Data":"6c041eb6552696acec4cdce751ffba6f4fadcbc11c748d4d9e94cb05149a48d4"} Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.229455 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.415130 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-util\") pod \"268ff739-d59c-4d7c-a413-bcb696ef8c86\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.415216 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqth2\" (UniqueName: \"kubernetes.io/projected/268ff739-d59c-4d7c-a413-bcb696ef8c86-kube-api-access-cqth2\") pod \"268ff739-d59c-4d7c-a413-bcb696ef8c86\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.415265 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-bundle\") pod \"268ff739-d59c-4d7c-a413-bcb696ef8c86\" (UID: \"268ff739-d59c-4d7c-a413-bcb696ef8c86\") " Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.416691 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-bundle" (OuterVolumeSpecName: "bundle") pod "268ff739-d59c-4d7c-a413-bcb696ef8c86" (UID: "268ff739-d59c-4d7c-a413-bcb696ef8c86"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.422435 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/268ff739-d59c-4d7c-a413-bcb696ef8c86-kube-api-access-cqth2" (OuterVolumeSpecName: "kube-api-access-cqth2") pod "268ff739-d59c-4d7c-a413-bcb696ef8c86" (UID: "268ff739-d59c-4d7c-a413-bcb696ef8c86"). InnerVolumeSpecName "kube-api-access-cqth2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.432381 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-util" (OuterVolumeSpecName: "util") pod "268ff739-d59c-4d7c-a413-bcb696ef8c86" (UID: "268ff739-d59c-4d7c-a413-bcb696ef8c86"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.517508 4946 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.517556 4946 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/268ff739-d59c-4d7c-a413-bcb696ef8c86-util\") on node \"crc\" DevicePath \"\"" Dec 03 07:07:26 crc kubenswrapper[4946]: I1203 07:07:26.517575 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqth2\" (UniqueName: \"kubernetes.io/projected/268ff739-d59c-4d7c-a413-bcb696ef8c86-kube-api-access-cqth2\") on node \"crc\" DevicePath \"\"" Dec 03 07:07:27 crc kubenswrapper[4946]: I1203 07:07:27.026159 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" event={"ID":"268ff739-d59c-4d7c-a413-bcb696ef8c86","Type":"ContainerDied","Data":"53a4686f428de83e7805dbecbfe5a89a0c91e762333a6828eeba07ca32ea9e84"} Dec 03 07:07:27 crc kubenswrapper[4946]: I1203 07:07:27.026565 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53a4686f428de83e7805dbecbfe5a89a0c91e762333a6828eeba07ca32ea9e84" Dec 03 07:07:27 crc kubenswrapper[4946]: I1203 07:07:27.026232 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.103580 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h"] Dec 03 07:07:33 crc kubenswrapper[4946]: E1203 07:07:33.104418 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerName="util" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.104433 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerName="util" Dec 03 07:07:33 crc kubenswrapper[4946]: E1203 07:07:33.104459 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerName="extract" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.104465 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerName="extract" Dec 03 07:07:33 crc kubenswrapper[4946]: E1203 07:07:33.104479 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerName="pull" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.104487 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerName="pull" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.104615 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="268ff739-d59c-4d7c-a413-bcb696ef8c86" containerName="extract" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.105083 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.108071 4946 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-4vfpp" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.112711 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.112972 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.119268 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h"] Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.206254 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqpw4\" (UniqueName: \"kubernetes.io/projected/4b8ebf2c-09a9-434f-b4d9-3b54b52b6610-kube-api-access-mqpw4\") pod \"cert-manager-operator-controller-manager-64cf6dff88-96z5h\" (UID: \"4b8ebf2c-09a9-434f-b4d9-3b54b52b6610\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.206396 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4b8ebf2c-09a9-434f-b4d9-3b54b52b6610-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-96z5h\" (UID: \"4b8ebf2c-09a9-434f-b4d9-3b54b52b6610\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.307684 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4b8ebf2c-09a9-434f-b4d9-3b54b52b6610-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-96z5h\" (UID: \"4b8ebf2c-09a9-434f-b4d9-3b54b52b6610\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.307771 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqpw4\" (UniqueName: \"kubernetes.io/projected/4b8ebf2c-09a9-434f-b4d9-3b54b52b6610-kube-api-access-mqpw4\") pod \"cert-manager-operator-controller-manager-64cf6dff88-96z5h\" (UID: \"4b8ebf2c-09a9-434f-b4d9-3b54b52b6610\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.308224 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4b8ebf2c-09a9-434f-b4d9-3b54b52b6610-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-96z5h\" (UID: \"4b8ebf2c-09a9-434f-b4d9-3b54b52b6610\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.327897 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqpw4\" (UniqueName: \"kubernetes.io/projected/4b8ebf2c-09a9-434f-b4d9-3b54b52b6610-kube-api-access-mqpw4\") pod \"cert-manager-operator-controller-manager-64cf6dff88-96z5h\" (UID: \"4b8ebf2c-09a9-434f-b4d9-3b54b52b6610\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.458666 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" Dec 03 07:07:33 crc kubenswrapper[4946]: I1203 07:07:33.714610 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h"] Dec 03 07:07:34 crc kubenswrapper[4946]: I1203 07:07:34.081327 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" event={"ID":"4b8ebf2c-09a9-434f-b4d9-3b54b52b6610","Type":"ContainerStarted","Data":"88781e758db977c188526430846e32db8ac6efa7cbd2938d63f01e47b1373b1c"} Dec 03 07:07:42 crc kubenswrapper[4946]: I1203 07:07:42.146211 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" event={"ID":"4b8ebf2c-09a9-434f-b4d9-3b54b52b6610","Type":"ContainerStarted","Data":"b87d4985ef010035f271ff65f1be28d80cf797cc4e112339ab873d8c72ef1f21"} Dec 03 07:07:42 crc kubenswrapper[4946]: I1203 07:07:42.164157 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-96z5h" podStartSLOduration=1.86290401 podStartE2EDuration="9.164141641s" podCreationTimestamp="2025-12-03 07:07:33 +0000 UTC" firstStartedPulling="2025-12-03 07:07:33.733331428 +0000 UTC m=+1046.530021537" lastFinishedPulling="2025-12-03 07:07:41.034569049 +0000 UTC m=+1053.831259168" observedRunningTime="2025-12-03 07:07:42.163144044 +0000 UTC m=+1054.959834153" watchObservedRunningTime="2025-12-03 07:07:42.164141641 +0000 UTC m=+1054.960831750" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.290204 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-wplrd"] Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.291891 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.294139 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.296171 4946 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-57j5f" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.296941 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.303410 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-wplrd"] Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.385173 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1837c927-ee2e-4452-aec6-9306ab8dff65-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-wplrd\" (UID: \"1837c927-ee2e-4452-aec6-9306ab8dff65\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.385484 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbrkx\" (UniqueName: \"kubernetes.io/projected/1837c927-ee2e-4452-aec6-9306ab8dff65-kube-api-access-mbrkx\") pod \"cert-manager-webhook-f4fb5df64-wplrd\" (UID: \"1837c927-ee2e-4452-aec6-9306ab8dff65\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.486177 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbrkx\" (UniqueName: \"kubernetes.io/projected/1837c927-ee2e-4452-aec6-9306ab8dff65-kube-api-access-mbrkx\") pod \"cert-manager-webhook-f4fb5df64-wplrd\" (UID: \"1837c927-ee2e-4452-aec6-9306ab8dff65\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.486254 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1837c927-ee2e-4452-aec6-9306ab8dff65-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-wplrd\" (UID: \"1837c927-ee2e-4452-aec6-9306ab8dff65\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.503856 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1837c927-ee2e-4452-aec6-9306ab8dff65-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-wplrd\" (UID: \"1837c927-ee2e-4452-aec6-9306ab8dff65\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.515042 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbrkx\" (UniqueName: \"kubernetes.io/projected/1837c927-ee2e-4452-aec6-9306ab8dff65-kube-api-access-mbrkx\") pod \"cert-manager-webhook-f4fb5df64-wplrd\" (UID: \"1837c927-ee2e-4452-aec6-9306ab8dff65\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:46 crc kubenswrapper[4946]: I1203 07:07:46.665178 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:47 crc kubenswrapper[4946]: I1203 07:07:47.104014 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-wplrd"] Dec 03 07:07:47 crc kubenswrapper[4946]: I1203 07:07:47.185374 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" event={"ID":"1837c927-ee2e-4452-aec6-9306ab8dff65","Type":"ContainerStarted","Data":"a2c981b96923dfdb3c0717af0f2d5eeaa72962a4218584a0ae7533cb0aac481e"} Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.019679 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-dcppn"] Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.020999 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.025972 4946 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-6t6lk" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.027769 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-dcppn"] Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.122286 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b6bc6f-2572-4b9b-99f2-c4ba512dcede-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-dcppn\" (UID: \"07b6bc6f-2572-4b9b-99f2-c4ba512dcede\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.122369 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2279\" (UniqueName: \"kubernetes.io/projected/07b6bc6f-2572-4b9b-99f2-c4ba512dcede-kube-api-access-x2279\") pod \"cert-manager-cainjector-855d9ccff4-dcppn\" (UID: \"07b6bc6f-2572-4b9b-99f2-c4ba512dcede\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.223255 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b6bc6f-2572-4b9b-99f2-c4ba512dcede-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-dcppn\" (UID: \"07b6bc6f-2572-4b9b-99f2-c4ba512dcede\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.223331 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2279\" (UniqueName: \"kubernetes.io/projected/07b6bc6f-2572-4b9b-99f2-c4ba512dcede-kube-api-access-x2279\") pod \"cert-manager-cainjector-855d9ccff4-dcppn\" (UID: \"07b6bc6f-2572-4b9b-99f2-c4ba512dcede\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.246273 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b6bc6f-2572-4b9b-99f2-c4ba512dcede-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-dcppn\" (UID: \"07b6bc6f-2572-4b9b-99f2-c4ba512dcede\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.247570 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2279\" (UniqueName: \"kubernetes.io/projected/07b6bc6f-2572-4b9b-99f2-c4ba512dcede-kube-api-access-x2279\") pod \"cert-manager-cainjector-855d9ccff4-dcppn\" (UID: \"07b6bc6f-2572-4b9b-99f2-c4ba512dcede\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.368089 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" Dec 03 07:07:49 crc kubenswrapper[4946]: I1203 07:07:49.822131 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-dcppn"] Dec 03 07:07:49 crc kubenswrapper[4946]: W1203 07:07:49.843902 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07b6bc6f_2572_4b9b_99f2_c4ba512dcede.slice/crio-301bd5c7e98053ce1bd6260bf96d708528b9d24948e22c2fb62dee2262a89e0d WatchSource:0}: Error finding container 301bd5c7e98053ce1bd6260bf96d708528b9d24948e22c2fb62dee2262a89e0d: Status 404 returned error can't find the container with id 301bd5c7e98053ce1bd6260bf96d708528b9d24948e22c2fb62dee2262a89e0d Dec 03 07:07:50 crc kubenswrapper[4946]: I1203 07:07:50.208306 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" event={"ID":"07b6bc6f-2572-4b9b-99f2-c4ba512dcede","Type":"ContainerStarted","Data":"301bd5c7e98053ce1bd6260bf96d708528b9d24948e22c2fb62dee2262a89e0d"} Dec 03 07:07:55 crc kubenswrapper[4946]: I1203 07:07:55.247443 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" event={"ID":"07b6bc6f-2572-4b9b-99f2-c4ba512dcede","Type":"ContainerStarted","Data":"560df8afc594fb8ab9bf65ba4d6a0678f74a2fa55d1edce4fe5caded8074a4a5"} Dec 03 07:07:55 crc kubenswrapper[4946]: I1203 07:07:55.250366 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" event={"ID":"1837c927-ee2e-4452-aec6-9306ab8dff65","Type":"ContainerStarted","Data":"46640f75f3cbb8d809c67f2d9718c8ea2cac9f6498a0de04aa45371b68f3719d"} Dec 03 07:07:55 crc kubenswrapper[4946]: I1203 07:07:55.250550 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:07:55 crc kubenswrapper[4946]: I1203 07:07:55.268126 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-dcppn" podStartSLOduration=1.367405328 podStartE2EDuration="6.268110501s" podCreationTimestamp="2025-12-03 07:07:49 +0000 UTC" firstStartedPulling="2025-12-03 07:07:49.850151212 +0000 UTC m=+1062.646841321" lastFinishedPulling="2025-12-03 07:07:54.750856385 +0000 UTC m=+1067.547546494" observedRunningTime="2025-12-03 07:07:55.265774538 +0000 UTC m=+1068.062464657" watchObservedRunningTime="2025-12-03 07:07:55.268110501 +0000 UTC m=+1068.064800630" Dec 03 07:07:55 crc kubenswrapper[4946]: I1203 07:07:55.293356 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" podStartSLOduration=1.831865383 podStartE2EDuration="9.29333724s" podCreationTimestamp="2025-12-03 07:07:46 +0000 UTC" firstStartedPulling="2025-12-03 07:07:47.113424281 +0000 UTC m=+1059.910114390" lastFinishedPulling="2025-12-03 07:07:54.574896138 +0000 UTC m=+1067.371586247" observedRunningTime="2025-12-03 07:07:55.292908628 +0000 UTC m=+1068.089598747" watchObservedRunningTime="2025-12-03 07:07:55.29333724 +0000 UTC m=+1068.090027359" Dec 03 07:08:01 crc kubenswrapper[4946]: I1203 07:08:01.670490 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-wplrd" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.226831 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-xcngx"] Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.227890 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-xcngx" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.230733 4946 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-9q8bq" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.239084 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-xcngx"] Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.379210 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnlb8\" (UniqueName: \"kubernetes.io/projected/babf0398-4ef7-4498-be2c-b970cc3db40c-kube-api-access-tnlb8\") pod \"cert-manager-86cb77c54b-xcngx\" (UID: \"babf0398-4ef7-4498-be2c-b970cc3db40c\") " pod="cert-manager/cert-manager-86cb77c54b-xcngx" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.379297 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/babf0398-4ef7-4498-be2c-b970cc3db40c-bound-sa-token\") pod \"cert-manager-86cb77c54b-xcngx\" (UID: \"babf0398-4ef7-4498-be2c-b970cc3db40c\") " pod="cert-manager/cert-manager-86cb77c54b-xcngx" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.480608 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/babf0398-4ef7-4498-be2c-b970cc3db40c-bound-sa-token\") pod \"cert-manager-86cb77c54b-xcngx\" (UID: \"babf0398-4ef7-4498-be2c-b970cc3db40c\") " pod="cert-manager/cert-manager-86cb77c54b-xcngx" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.480839 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnlb8\" (UniqueName: \"kubernetes.io/projected/babf0398-4ef7-4498-be2c-b970cc3db40c-kube-api-access-tnlb8\") pod \"cert-manager-86cb77c54b-xcngx\" (UID: \"babf0398-4ef7-4498-be2c-b970cc3db40c\") " pod="cert-manager/cert-manager-86cb77c54b-xcngx" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.501216 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/babf0398-4ef7-4498-be2c-b970cc3db40c-bound-sa-token\") pod \"cert-manager-86cb77c54b-xcngx\" (UID: \"babf0398-4ef7-4498-be2c-b970cc3db40c\") " pod="cert-manager/cert-manager-86cb77c54b-xcngx" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.510257 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnlb8\" (UniqueName: \"kubernetes.io/projected/babf0398-4ef7-4498-be2c-b970cc3db40c-kube-api-access-tnlb8\") pod \"cert-manager-86cb77c54b-xcngx\" (UID: \"babf0398-4ef7-4498-be2c-b970cc3db40c\") " pod="cert-manager/cert-manager-86cb77c54b-xcngx" Dec 03 07:08:04 crc kubenswrapper[4946]: I1203 07:08:04.555808 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-xcngx" Dec 03 07:08:06 crc kubenswrapper[4946]: I1203 07:08:06.286586 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-xcngx"] Dec 03 07:08:06 crc kubenswrapper[4946]: I1203 07:08:06.326231 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-xcngx" event={"ID":"babf0398-4ef7-4498-be2c-b970cc3db40c","Type":"ContainerStarted","Data":"2dc2d0152230c6e3e10dc55d00ee2b312953fbbf9f0d92b813b9c84a1b1859a9"} Dec 03 07:08:09 crc kubenswrapper[4946]: I1203 07:08:09.349601 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-xcngx" event={"ID":"babf0398-4ef7-4498-be2c-b970cc3db40c","Type":"ContainerStarted","Data":"3c03bd7df3cf4a88644d50b8502a3818a07118d9a2b0184d5b794fea508e632a"} Dec 03 07:08:09 crc kubenswrapper[4946]: I1203 07:08:09.384126 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-xcngx" podStartSLOduration=5.384096018 podStartE2EDuration="5.384096018s" podCreationTimestamp="2025-12-03 07:08:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:08:09.374362196 +0000 UTC m=+1082.171052365" watchObservedRunningTime="2025-12-03 07:08:09.384096018 +0000 UTC m=+1082.180786157" Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.711026 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rmbpq"] Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.712818 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rmbpq" Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.716195 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-g9s2c" Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.716942 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.716951 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.724877 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rmbpq"] Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.848602 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/f6eafd8b-d5ab-4e24-a978-6a299649be03-kube-api-access-djwpk\") pod \"openstack-operator-index-rmbpq\" (UID: \"f6eafd8b-d5ab-4e24-a978-6a299649be03\") " pod="openstack-operators/openstack-operator-index-rmbpq" Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.950382 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/f6eafd8b-d5ab-4e24-a978-6a299649be03-kube-api-access-djwpk\") pod \"openstack-operator-index-rmbpq\" (UID: \"f6eafd8b-d5ab-4e24-a978-6a299649be03\") " pod="openstack-operators/openstack-operator-index-rmbpq" Dec 03 07:08:15 crc kubenswrapper[4946]: I1203 07:08:15.971193 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/f6eafd8b-d5ab-4e24-a978-6a299649be03-kube-api-access-djwpk\") pod \"openstack-operator-index-rmbpq\" (UID: \"f6eafd8b-d5ab-4e24-a978-6a299649be03\") " pod="openstack-operators/openstack-operator-index-rmbpq" Dec 03 07:08:16 crc kubenswrapper[4946]: I1203 07:08:16.048089 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rmbpq" Dec 03 07:08:16 crc kubenswrapper[4946]: I1203 07:08:16.359247 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rmbpq"] Dec 03 07:08:16 crc kubenswrapper[4946]: I1203 07:08:16.398772 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rmbpq" event={"ID":"f6eafd8b-d5ab-4e24-a978-6a299649be03","Type":"ContainerStarted","Data":"bcc3c6de3f7fee8683ee6c3b1acd883783b76a4dcaf4e06b95cb0f9e88a53958"} Dec 03 07:08:17 crc kubenswrapper[4946]: I1203 07:08:17.410579 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rmbpq" event={"ID":"f6eafd8b-d5ab-4e24-a978-6a299649be03","Type":"ContainerStarted","Data":"52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4"} Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.043924 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rmbpq" podStartSLOduration=3.333487372 podStartE2EDuration="4.043899567s" podCreationTimestamp="2025-12-03 07:08:15 +0000 UTC" firstStartedPulling="2025-12-03 07:08:16.365940829 +0000 UTC m=+1089.162630938" lastFinishedPulling="2025-12-03 07:08:17.076353024 +0000 UTC m=+1089.873043133" observedRunningTime="2025-12-03 07:08:17.431167417 +0000 UTC m=+1090.227857566" watchObservedRunningTime="2025-12-03 07:08:19.043899567 +0000 UTC m=+1091.840589716" Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.049839 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rmbpq"] Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.427250 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-rmbpq" podUID="f6eafd8b-d5ab-4e24-a978-6a299649be03" containerName="registry-server" containerID="cri-o://52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4" gracePeriod=2 Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.655238 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fq7s4"] Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.656229 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.671053 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fq7s4"] Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.808689 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jncj9\" (UniqueName: \"kubernetes.io/projected/1b592596-16d1-410d-953b-ad7c04d2f6a4-kube-api-access-jncj9\") pod \"openstack-operator-index-fq7s4\" (UID: \"1b592596-16d1-410d-953b-ad7c04d2f6a4\") " pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.823207 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rmbpq" Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.909783 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jncj9\" (UniqueName: \"kubernetes.io/projected/1b592596-16d1-410d-953b-ad7c04d2f6a4-kube-api-access-jncj9\") pod \"openstack-operator-index-fq7s4\" (UID: \"1b592596-16d1-410d-953b-ad7c04d2f6a4\") " pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.929180 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jncj9\" (UniqueName: \"kubernetes.io/projected/1b592596-16d1-410d-953b-ad7c04d2f6a4-kube-api-access-jncj9\") pod \"openstack-operator-index-fq7s4\" (UID: \"1b592596-16d1-410d-953b-ad7c04d2f6a4\") " pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:19 crc kubenswrapper[4946]: I1203 07:08:19.986466 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.010728 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/f6eafd8b-d5ab-4e24-a978-6a299649be03-kube-api-access-djwpk\") pod \"f6eafd8b-d5ab-4e24-a978-6a299649be03\" (UID: \"f6eafd8b-d5ab-4e24-a978-6a299649be03\") " Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.013339 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6eafd8b-d5ab-4e24-a978-6a299649be03-kube-api-access-djwpk" (OuterVolumeSpecName: "kube-api-access-djwpk") pod "f6eafd8b-d5ab-4e24-a978-6a299649be03" (UID: "f6eafd8b-d5ab-4e24-a978-6a299649be03"). InnerVolumeSpecName "kube-api-access-djwpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.112355 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/f6eafd8b-d5ab-4e24-a978-6a299649be03-kube-api-access-djwpk\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.229117 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fq7s4"] Dec 03 07:08:20 crc kubenswrapper[4946]: W1203 07:08:20.233325 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b592596_16d1_410d_953b_ad7c04d2f6a4.slice/crio-babd3f8b8e42de2116bafb52cebe5ff9d486fa392d427fdd684ee02ec9f9c4a8 WatchSource:0}: Error finding container babd3f8b8e42de2116bafb52cebe5ff9d486fa392d427fdd684ee02ec9f9c4a8: Status 404 returned error can't find the container with id babd3f8b8e42de2116bafb52cebe5ff9d486fa392d427fdd684ee02ec9f9c4a8 Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.436601 4946 generic.go:334] "Generic (PLEG): container finished" podID="f6eafd8b-d5ab-4e24-a978-6a299649be03" containerID="52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4" exitCode=0 Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.436715 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rmbpq" Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.436959 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rmbpq" event={"ID":"f6eafd8b-d5ab-4e24-a978-6a299649be03","Type":"ContainerDied","Data":"52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4"} Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.436990 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rmbpq" event={"ID":"f6eafd8b-d5ab-4e24-a978-6a299649be03","Type":"ContainerDied","Data":"bcc3c6de3f7fee8683ee6c3b1acd883783b76a4dcaf4e06b95cb0f9e88a53958"} Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.437007 4946 scope.go:117] "RemoveContainer" containerID="52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4" Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.441860 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fq7s4" event={"ID":"1b592596-16d1-410d-953b-ad7c04d2f6a4","Type":"ContainerStarted","Data":"babd3f8b8e42de2116bafb52cebe5ff9d486fa392d427fdd684ee02ec9f9c4a8"} Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.475653 4946 scope.go:117] "RemoveContainer" containerID="52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4" Dec 03 07:08:20 crc kubenswrapper[4946]: E1203 07:08:20.481255 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4\": container with ID starting with 52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4 not found: ID does not exist" containerID="52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4" Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.481321 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4"} err="failed to get container status \"52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4\": rpc error: code = NotFound desc = could not find container \"52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4\": container with ID starting with 52fad197c9b8a1d8fa463ee6940ddc2aeb8f6b02fde2a1223604fb00830d1da4 not found: ID does not exist" Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.484853 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rmbpq"] Dec 03 07:08:20 crc kubenswrapper[4946]: I1203 07:08:20.490482 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-rmbpq"] Dec 03 07:08:21 crc kubenswrapper[4946]: I1203 07:08:21.454211 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fq7s4" event={"ID":"1b592596-16d1-410d-953b-ad7c04d2f6a4","Type":"ContainerStarted","Data":"cdfba5eb2569235ee6b312ca37aad3db205459c7f552d19dce1ed3093b312d4a"} Dec 03 07:08:21 crc kubenswrapper[4946]: I1203 07:08:21.487230 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fq7s4" podStartSLOduration=2.049880205 podStartE2EDuration="2.487210288s" podCreationTimestamp="2025-12-03 07:08:19 +0000 UTC" firstStartedPulling="2025-12-03 07:08:20.240245827 +0000 UTC m=+1093.036935946" lastFinishedPulling="2025-12-03 07:08:20.67757591 +0000 UTC m=+1093.474266029" observedRunningTime="2025-12-03 07:08:21.477297311 +0000 UTC m=+1094.273987460" watchObservedRunningTime="2025-12-03 07:08:21.487210288 +0000 UTC m=+1094.283900417" Dec 03 07:08:21 crc kubenswrapper[4946]: I1203 07:08:21.600246 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6eafd8b-d5ab-4e24-a978-6a299649be03" path="/var/lib/kubelet/pods/f6eafd8b-d5ab-4e24-a978-6a299649be03/volumes" Dec 03 07:08:23 crc kubenswrapper[4946]: I1203 07:08:23.039184 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:08:23 crc kubenswrapper[4946]: I1203 07:08:23.040423 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:08:29 crc kubenswrapper[4946]: I1203 07:08:29.986687 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:29 crc kubenswrapper[4946]: I1203 07:08:29.987265 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:30 crc kubenswrapper[4946]: I1203 07:08:30.024027 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:30 crc kubenswrapper[4946]: I1203 07:08:30.572308 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-fq7s4" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.696412 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p"] Dec 03 07:08:36 crc kubenswrapper[4946]: E1203 07:08:36.697269 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6eafd8b-d5ab-4e24-a978-6a299649be03" containerName="registry-server" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.697284 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6eafd8b-d5ab-4e24-a978-6a299649be03" containerName="registry-server" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.697440 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6eafd8b-d5ab-4e24-a978-6a299649be03" containerName="registry-server" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.698418 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.704278 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-v47f2" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.708321 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p"] Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.854790 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-bundle\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.854955 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx226\" (UniqueName: \"kubernetes.io/projected/e21599fd-a0aa-46eb-80d9-a50c426bb629-kube-api-access-hx226\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.855049 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-util\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.957092 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx226\" (UniqueName: \"kubernetes.io/projected/e21599fd-a0aa-46eb-80d9-a50c426bb629-kube-api-access-hx226\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.957239 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-util\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.957330 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-bundle\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.958280 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-bundle\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.958305 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-util\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:36 crc kubenswrapper[4946]: I1203 07:08:36.983824 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx226\" (UniqueName: \"kubernetes.io/projected/e21599fd-a0aa-46eb-80d9-a50c426bb629-kube-api-access-hx226\") pod \"98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:37 crc kubenswrapper[4946]: I1203 07:08:37.020269 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:37 crc kubenswrapper[4946]: I1203 07:08:37.238997 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p"] Dec 03 07:08:37 crc kubenswrapper[4946]: I1203 07:08:37.592299 4946 generic.go:334] "Generic (PLEG): container finished" podID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerID="64b50561823bd7fb4535187e840ea35bec24f49902d199aee243b8e707478466" exitCode=0 Dec 03 07:08:37 crc kubenswrapper[4946]: I1203 07:08:37.600237 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:08:37 crc kubenswrapper[4946]: I1203 07:08:37.601088 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" event={"ID":"e21599fd-a0aa-46eb-80d9-a50c426bb629","Type":"ContainerDied","Data":"64b50561823bd7fb4535187e840ea35bec24f49902d199aee243b8e707478466"} Dec 03 07:08:37 crc kubenswrapper[4946]: I1203 07:08:37.601144 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" event={"ID":"e21599fd-a0aa-46eb-80d9-a50c426bb629","Type":"ContainerStarted","Data":"cfbf6d962a8b43126325f85942e26d17fba34447f9e6019be9ac58276d7a6c16"} Dec 03 07:08:38 crc kubenswrapper[4946]: I1203 07:08:38.608227 4946 generic.go:334] "Generic (PLEG): container finished" podID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerID="0e84731d6c99222da4f9a62a0ac1e681cd3cab0a8499fd069ccfb177259b0d73" exitCode=0 Dec 03 07:08:38 crc kubenswrapper[4946]: I1203 07:08:38.608281 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" event={"ID":"e21599fd-a0aa-46eb-80d9-a50c426bb629","Type":"ContainerDied","Data":"0e84731d6c99222da4f9a62a0ac1e681cd3cab0a8499fd069ccfb177259b0d73"} Dec 03 07:08:39 crc kubenswrapper[4946]: I1203 07:08:39.618082 4946 generic.go:334] "Generic (PLEG): container finished" podID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerID="8f94a9bdcf803b43f79ab34b0c936ca87ca92cd6f50400978bf8ab89237bdc8e" exitCode=0 Dec 03 07:08:39 crc kubenswrapper[4946]: I1203 07:08:39.618200 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" event={"ID":"e21599fd-a0aa-46eb-80d9-a50c426bb629","Type":"ContainerDied","Data":"8f94a9bdcf803b43f79ab34b0c936ca87ca92cd6f50400978bf8ab89237bdc8e"} Dec 03 07:08:40 crc kubenswrapper[4946]: I1203 07:08:40.944535 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.015553 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-bundle\") pod \"e21599fd-a0aa-46eb-80d9-a50c426bb629\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.015704 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx226\" (UniqueName: \"kubernetes.io/projected/e21599fd-a0aa-46eb-80d9-a50c426bb629-kube-api-access-hx226\") pod \"e21599fd-a0aa-46eb-80d9-a50c426bb629\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.015802 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-util\") pod \"e21599fd-a0aa-46eb-80d9-a50c426bb629\" (UID: \"e21599fd-a0aa-46eb-80d9-a50c426bb629\") " Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.016471 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-bundle" (OuterVolumeSpecName: "bundle") pod "e21599fd-a0aa-46eb-80d9-a50c426bb629" (UID: "e21599fd-a0aa-46eb-80d9-a50c426bb629"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.021669 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e21599fd-a0aa-46eb-80d9-a50c426bb629-kube-api-access-hx226" (OuterVolumeSpecName: "kube-api-access-hx226") pod "e21599fd-a0aa-46eb-80d9-a50c426bb629" (UID: "e21599fd-a0aa-46eb-80d9-a50c426bb629"). InnerVolumeSpecName "kube-api-access-hx226". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.039571 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-util" (OuterVolumeSpecName: "util") pod "e21599fd-a0aa-46eb-80d9-a50c426bb629" (UID: "e21599fd-a0aa-46eb-80d9-a50c426bb629"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.118040 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx226\" (UniqueName: \"kubernetes.io/projected/e21599fd-a0aa-46eb-80d9-a50c426bb629-kube-api-access-hx226\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.118116 4946 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-util\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.118145 4946 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e21599fd-a0aa-46eb-80d9-a50c426bb629-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.638837 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" event={"ID":"e21599fd-a0aa-46eb-80d9-a50c426bb629","Type":"ContainerDied","Data":"cfbf6d962a8b43126325f85942e26d17fba34447f9e6019be9ac58276d7a6c16"} Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.638871 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfbf6d962a8b43126325f85942e26d17fba34447f9e6019be9ac58276d7a6c16" Dec 03 07:08:41 crc kubenswrapper[4946]: I1203 07:08:41.638931 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p" Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.885656 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp"] Dec 03 07:08:44 crc kubenswrapper[4946]: E1203 07:08:44.888258 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerName="pull" Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.888482 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerName="pull" Dec 03 07:08:44 crc kubenswrapper[4946]: E1203 07:08:44.888667 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerName="extract" Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.888896 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerName="extract" Dec 03 07:08:44 crc kubenswrapper[4946]: E1203 07:08:44.889053 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerName="util" Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.889202 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerName="util" Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.889630 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="e21599fd-a0aa-46eb-80d9-a50c426bb629" containerName="extract" Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.890685 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.893689 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-jgx6d" Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.922309 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp"] Dec 03 07:08:44 crc kubenswrapper[4946]: I1203 07:08:44.969344 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtm8g\" (UniqueName: \"kubernetes.io/projected/9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77-kube-api-access-rtm8g\") pod \"openstack-operator-controller-operator-7dd5c7bb7c-hh9qp\" (UID: \"9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77\") " pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" Dec 03 07:08:45 crc kubenswrapper[4946]: I1203 07:08:45.070512 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtm8g\" (UniqueName: \"kubernetes.io/projected/9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77-kube-api-access-rtm8g\") pod \"openstack-operator-controller-operator-7dd5c7bb7c-hh9qp\" (UID: \"9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77\") " pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" Dec 03 07:08:45 crc kubenswrapper[4946]: I1203 07:08:45.097220 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtm8g\" (UniqueName: \"kubernetes.io/projected/9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77-kube-api-access-rtm8g\") pod \"openstack-operator-controller-operator-7dd5c7bb7c-hh9qp\" (UID: \"9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77\") " pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" Dec 03 07:08:45 crc kubenswrapper[4946]: I1203 07:08:45.215314 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" Dec 03 07:08:45 crc kubenswrapper[4946]: I1203 07:08:45.727293 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp"] Dec 03 07:08:46 crc kubenswrapper[4946]: I1203 07:08:46.671624 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" event={"ID":"9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77","Type":"ContainerStarted","Data":"770a0a7e7915490f3d4a10486d81263ac5ca53f6955b67794e1ae8c4dfa34010"} Dec 03 07:08:50 crc kubenswrapper[4946]: I1203 07:08:50.705146 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" event={"ID":"9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77","Type":"ContainerStarted","Data":"6fbb56ec68481d013bb8fae5ba18b9be22001688b8cd0107c3ce8b93b79f291d"} Dec 03 07:08:50 crc kubenswrapper[4946]: I1203 07:08:50.706003 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" Dec 03 07:08:50 crc kubenswrapper[4946]: I1203 07:08:50.762865 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" podStartSLOduration=2.799074803 podStartE2EDuration="6.762843009s" podCreationTimestamp="2025-12-03 07:08:44 +0000 UTC" firstStartedPulling="2025-12-03 07:08:45.735862823 +0000 UTC m=+1118.532552982" lastFinishedPulling="2025-12-03 07:08:49.699631069 +0000 UTC m=+1122.496321188" observedRunningTime="2025-12-03 07:08:50.758975856 +0000 UTC m=+1123.555666045" watchObservedRunningTime="2025-12-03 07:08:50.762843009 +0000 UTC m=+1123.559533148" Dec 03 07:08:53 crc kubenswrapper[4946]: I1203 07:08:53.039300 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:08:53 crc kubenswrapper[4946]: I1203 07:08:53.039693 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:08:55 crc kubenswrapper[4946]: I1203 07:08:55.219621 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7dd5c7bb7c-hh9qp" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.547588 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.549164 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.551265 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.552183 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.552649 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-ndzmg" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.554870 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-bcz26" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.563396 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.567820 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.582493 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.583470 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.588212 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.589337 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.602533 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-qzk6l" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.609924 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-jr5kz" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.610941 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.612089 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.614322 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-9twjz" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.618452 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.631635 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.644636 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8tv6\" (UniqueName: \"kubernetes.io/projected/881fe69d-7e12-4250-8589-76096d3318d9-kube-api-access-p8tv6\") pod \"barbican-operator-controller-manager-7d9dfd778-v62hf\" (UID: \"881fe69d-7e12-4250-8589-76096d3318d9\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.644781 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh96g\" (UniqueName: \"kubernetes.io/projected/592abc1b-947b-4b47-965f-0110fa5e74e3-kube-api-access-fh96g\") pod \"cinder-operator-controller-manager-859b6ccc6-v7djg\" (UID: \"592abc1b-947b-4b47-965f-0110fa5e74e3\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.644803 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jrqt\" (UniqueName: \"kubernetes.io/projected/feaece87-4f9e-4945-8faa-c49487a57c2f-kube-api-access-6jrqt\") pod \"designate-operator-controller-manager-78b4bc895b-fwbh4\" (UID: \"feaece87-4f9e-4945-8faa-c49487a57c2f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.644823 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwgrx\" (UniqueName: \"kubernetes.io/projected/f050d6c8-74b6-42cb-943b-6e372c3e0dfe-kube-api-access-wwgrx\") pod \"glance-operator-controller-manager-77987cd8cd-rkv2m\" (UID: \"f050d6c8-74b6-42cb-943b-6e372c3e0dfe\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.644866 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4cj6\" (UniqueName: \"kubernetes.io/projected/afb5f2e7-fb0a-425b-9a3c-cff10fdec757-kube-api-access-k4cj6\") pod \"heat-operator-controller-manager-5f64f6f8bb-7cwn9\" (UID: \"afb5f2e7-fb0a-425b-9a3c-cff10fdec757\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.653827 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.654889 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.656824 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-2hthl" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.677818 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.689129 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.694660 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-lvk95"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.695567 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.697849 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-cg8bn" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.698564 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.717090 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-lvk95"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.725453 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.726552 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.732153 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-6mwqk" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.745997 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7s28\" (UniqueName: \"kubernetes.io/projected/bd12d71a-063b-445c-876b-aab60ef74f35-kube-api-access-r7s28\") pod \"horizon-operator-controller-manager-68c6d99b8f-sv727\" (UID: \"bd12d71a-063b-445c-876b-aab60ef74f35\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.746045 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4cj6\" (UniqueName: \"kubernetes.io/projected/afb5f2e7-fb0a-425b-9a3c-cff10fdec757-kube-api-access-k4cj6\") pod \"heat-operator-controller-manager-5f64f6f8bb-7cwn9\" (UID: \"afb5f2e7-fb0a-425b-9a3c-cff10fdec757\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.746095 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8tv6\" (UniqueName: \"kubernetes.io/projected/881fe69d-7e12-4250-8589-76096d3318d9-kube-api-access-p8tv6\") pod \"barbican-operator-controller-manager-7d9dfd778-v62hf\" (UID: \"881fe69d-7e12-4250-8589-76096d3318d9\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.746120 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gtr4\" (UniqueName: \"kubernetes.io/projected/26f08bfc-aab0-4d18-a223-8e93309c15b0-kube-api-access-9gtr4\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.746141 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.746171 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh96g\" (UniqueName: \"kubernetes.io/projected/592abc1b-947b-4b47-965f-0110fa5e74e3-kube-api-access-fh96g\") pod \"cinder-operator-controller-manager-859b6ccc6-v7djg\" (UID: \"592abc1b-947b-4b47-965f-0110fa5e74e3\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.746189 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jrqt\" (UniqueName: \"kubernetes.io/projected/feaece87-4f9e-4945-8faa-c49487a57c2f-kube-api-access-6jrqt\") pod \"designate-operator-controller-manager-78b4bc895b-fwbh4\" (UID: \"feaece87-4f9e-4945-8faa-c49487a57c2f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.746207 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dv2d\" (UniqueName: \"kubernetes.io/projected/29fe18fb-39d7-4003-8171-0e47aa1d1b18-kube-api-access-7dv2d\") pod \"ironic-operator-controller-manager-6c548fd776-q88nf\" (UID: \"29fe18fb-39d7-4003-8171-0e47aa1d1b18\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.746225 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwgrx\" (UniqueName: \"kubernetes.io/projected/f050d6c8-74b6-42cb-943b-6e372c3e0dfe-kube-api-access-wwgrx\") pod \"glance-operator-controller-manager-77987cd8cd-rkv2m\" (UID: \"f050d6c8-74b6-42cb-943b-6e372c3e0dfe\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.753518 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.754894 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.768012 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.772058 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-x4s59" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.782827 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jrqt\" (UniqueName: \"kubernetes.io/projected/feaece87-4f9e-4945-8faa-c49487a57c2f-kube-api-access-6jrqt\") pod \"designate-operator-controller-manager-78b4bc895b-fwbh4\" (UID: \"feaece87-4f9e-4945-8faa-c49487a57c2f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.789368 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.790303 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.790341 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh96g\" (UniqueName: \"kubernetes.io/projected/592abc1b-947b-4b47-965f-0110fa5e74e3-kube-api-access-fh96g\") pod \"cinder-operator-controller-manager-859b6ccc6-v7djg\" (UID: \"592abc1b-947b-4b47-965f-0110fa5e74e3\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.791390 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8tv6\" (UniqueName: \"kubernetes.io/projected/881fe69d-7e12-4250-8589-76096d3318d9-kube-api-access-p8tv6\") pod \"barbican-operator-controller-manager-7d9dfd778-v62hf\" (UID: \"881fe69d-7e12-4250-8589-76096d3318d9\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.792146 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwgrx\" (UniqueName: \"kubernetes.io/projected/f050d6c8-74b6-42cb-943b-6e372c3e0dfe-kube-api-access-wwgrx\") pod \"glance-operator-controller-manager-77987cd8cd-rkv2m\" (UID: \"f050d6c8-74b6-42cb-943b-6e372c3e0dfe\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.794240 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4cj6\" (UniqueName: \"kubernetes.io/projected/afb5f2e7-fb0a-425b-9a3c-cff10fdec757-kube-api-access-k4cj6\") pod \"heat-operator-controller-manager-5f64f6f8bb-7cwn9\" (UID: \"afb5f2e7-fb0a-425b-9a3c-cff10fdec757\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.800356 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.805137 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.806163 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.812934 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8tlxp" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.813015 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-5drsz" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.833125 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.848856 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gtr4\" (UniqueName: \"kubernetes.io/projected/26f08bfc-aab0-4d18-a223-8e93309c15b0-kube-api-access-9gtr4\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.849606 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.849663 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dv2d\" (UniqueName: \"kubernetes.io/projected/29fe18fb-39d7-4003-8171-0e47aa1d1b18-kube-api-access-7dv2d\") pod \"ironic-operator-controller-manager-6c548fd776-q88nf\" (UID: \"29fe18fb-39d7-4003-8171-0e47aa1d1b18\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.849705 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7s28\" (UniqueName: \"kubernetes.io/projected/bd12d71a-063b-445c-876b-aab60ef74f35-kube-api-access-r7s28\") pod \"horizon-operator-controller-manager-68c6d99b8f-sv727\" (UID: \"bd12d71a-063b-445c-876b-aab60ef74f35\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" Dec 03 07:09:13 crc kubenswrapper[4946]: E1203 07:09:13.849873 4946 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:13 crc kubenswrapper[4946]: E1203 07:09:13.849956 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert podName:26f08bfc-aab0-4d18-a223-8e93309c15b0 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:14.349937658 +0000 UTC m=+1147.146627767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert") pod "infra-operator-controller-manager-57548d458d-lvk95" (UID: "26f08bfc-aab0-4d18-a223-8e93309c15b0") : secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.866071 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.875487 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.883303 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.897499 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7s28\" (UniqueName: \"kubernetes.io/projected/bd12d71a-063b-445c-876b-aab60ef74f35-kube-api-access-r7s28\") pod \"horizon-operator-controller-manager-68c6d99b8f-sv727\" (UID: \"bd12d71a-063b-445c-876b-aab60ef74f35\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.897557 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.898814 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.901214 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.905310 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gtr4\" (UniqueName: \"kubernetes.io/projected/26f08bfc-aab0-4d18-a223-8e93309c15b0-kube-api-access-9gtr4\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.901985 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-kx2sl" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.918939 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.919165 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.929207 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.935700 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.947590 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.947881 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.951479 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-ljq8v" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.952872 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsltj\" (UniqueName: \"kubernetes.io/projected/9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81-kube-api-access-zsltj\") pod \"manila-operator-controller-manager-7c79b5df47-q27df\" (UID: \"9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.952963 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8z9q\" (UniqueName: \"kubernetes.io/projected/4110947d-509f-4bb7-bb92-e52ebb362340-kube-api-access-k8z9q\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2hnpw\" (UID: \"4110947d-509f-4bb7-bb92-e52ebb362340\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.953005 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn9qf\" (UniqueName: \"kubernetes.io/projected/f55d3e9f-d953-4672-9438-4f8db6a0af1b-kube-api-access-fn9qf\") pod \"keystone-operator-controller-manager-7765d96ddf-b68fp\" (UID: \"f55d3e9f-d953-4672-9438-4f8db6a0af1b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.954901 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.966541 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.972190 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.972277 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-5hrj4" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.975428 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.982495 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dv2d\" (UniqueName: \"kubernetes.io/projected/29fe18fb-39d7-4003-8171-0e47aa1d1b18-kube-api-access-7dv2d\") pod \"ironic-operator-controller-manager-6c548fd776-q88nf\" (UID: \"29fe18fb-39d7-4003-8171-0e47aa1d1b18\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.982901 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.984303 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.985520 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.986205 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-skl42" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.989796 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c"] Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.991385 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.993053 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5k2dl" Dec 03 07:09:13 crc kubenswrapper[4946]: I1203 07:09:13.994281 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.013402 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-p28rk"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.015330 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.018671 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-f8vhb" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.025816 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.032788 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-p28rk"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.053658 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.054828 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056682 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhlqm\" (UniqueName: \"kubernetes.io/projected/ab0a2041-fd3d-49dc-b184-fdd603df844c-kube-api-access-mhlqm\") pod \"swift-operator-controller-manager-5f8c65bbfc-w9prn\" (UID: \"ab0a2041-fd3d-49dc-b184-fdd603df844c\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056727 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056763 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsltj\" (UniqueName: \"kubernetes.io/projected/9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81-kube-api-access-zsltj\") pod \"manila-operator-controller-manager-7c79b5df47-q27df\" (UID: \"9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056784 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmvs9\" (UniqueName: \"kubernetes.io/projected/06dff8f3-1b1f-4905-8976-ad14fd8b0b3f-kube-api-access-xmvs9\") pod \"ovn-operator-controller-manager-b6456fdb6-zws5c\" (UID: \"06dff8f3-1b1f-4905-8976-ad14fd8b0b3f\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056828 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8z9q\" (UniqueName: \"kubernetes.io/projected/4110947d-509f-4bb7-bb92-e52ebb362340-kube-api-access-k8z9q\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2hnpw\" (UID: \"4110947d-509f-4bb7-bb92-e52ebb362340\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056849 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm99k\" (UniqueName: \"kubernetes.io/projected/77e8b06b-4c6d-478a-ac26-230010446011-kube-api-access-hm99k\") pod \"octavia-operator-controller-manager-998648c74-xlsd9\" (UID: \"77e8b06b-4c6d-478a-ac26-230010446011\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056872 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn9qf\" (UniqueName: \"kubernetes.io/projected/f55d3e9f-d953-4672-9438-4f8db6a0af1b-kube-api-access-fn9qf\") pod \"keystone-operator-controller-manager-7765d96ddf-b68fp\" (UID: \"f55d3e9f-d953-4672-9438-4f8db6a0af1b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056891 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf6df\" (UniqueName: \"kubernetes.io/projected/f6febc9c-4437-4f94-8996-f4ece33a42f2-kube-api-access-vf6df\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-9n79w\" (UID: \"f6febc9c-4437-4f94-8996-f4ece33a42f2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056910 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ff7f\" (UniqueName: \"kubernetes.io/projected/4bbd7e74-3c10-47ea-ac92-2845c974a128-kube-api-access-5ff7f\") pod \"placement-operator-controller-manager-78f8948974-p28rk\" (UID: \"4bbd7e74-3c10-47ea-ac92-2845c974a128\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056927 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncqlp\" (UniqueName: \"kubernetes.io/projected/59735e89-ef8b-4887-9af2-d2bd5d3f7bb0-kube-api-access-ncqlp\") pod \"nova-operator-controller-manager-697bc559fc-g6cf2\" (UID: \"59735e89-ef8b-4887-9af2-d2bd5d3f7bb0\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.056954 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrrh6\" (UniqueName: \"kubernetes.io/projected/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-kube-api-access-zrrh6\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.057709 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-mjfct" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.065975 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.076226 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.093837 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.094985 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.097446 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8z9q\" (UniqueName: \"kubernetes.io/projected/4110947d-509f-4bb7-bb92-e52ebb362340-kube-api-access-k8z9q\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2hnpw\" (UID: \"4110947d-509f-4bb7-bb92-e52ebb362340\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.098379 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-hfvc6" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.100763 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsltj\" (UniqueName: \"kubernetes.io/projected/9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81-kube-api-access-zsltj\") pod \"manila-operator-controller-manager-7c79b5df47-q27df\" (UID: \"9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.104003 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.112345 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn9qf\" (UniqueName: \"kubernetes.io/projected/f55d3e9f-d953-4672-9438-4f8db6a0af1b-kube-api-access-fn9qf\") pod \"keystone-operator-controller-manager-7765d96ddf-b68fp\" (UID: \"f55d3e9f-d953-4672-9438-4f8db6a0af1b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.141164 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.144568 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.145649 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.161125 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm99k\" (UniqueName: \"kubernetes.io/projected/77e8b06b-4c6d-478a-ac26-230010446011-kube-api-access-hm99k\") pod \"octavia-operator-controller-manager-998648c74-xlsd9\" (UID: \"77e8b06b-4c6d-478a-ac26-230010446011\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.161167 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf6df\" (UniqueName: \"kubernetes.io/projected/f6febc9c-4437-4f94-8996-f4ece33a42f2-kube-api-access-vf6df\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-9n79w\" (UID: \"f6febc9c-4437-4f94-8996-f4ece33a42f2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.161188 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ff7f\" (UniqueName: \"kubernetes.io/projected/4bbd7e74-3c10-47ea-ac92-2845c974a128-kube-api-access-5ff7f\") pod \"placement-operator-controller-manager-78f8948974-p28rk\" (UID: \"4bbd7e74-3c10-47ea-ac92-2845c974a128\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.161208 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncqlp\" (UniqueName: \"kubernetes.io/projected/59735e89-ef8b-4887-9af2-d2bd5d3f7bb0-kube-api-access-ncqlp\") pod \"nova-operator-controller-manager-697bc559fc-g6cf2\" (UID: \"59735e89-ef8b-4887-9af2-d2bd5d3f7bb0\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.161232 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrrh6\" (UniqueName: \"kubernetes.io/projected/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-kube-api-access-zrrh6\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.161262 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhlqm\" (UniqueName: \"kubernetes.io/projected/ab0a2041-fd3d-49dc-b184-fdd603df844c-kube-api-access-mhlqm\") pod \"swift-operator-controller-manager-5f8c65bbfc-w9prn\" (UID: \"ab0a2041-fd3d-49dc-b184-fdd603df844c\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.161292 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.161313 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmvs9\" (UniqueName: \"kubernetes.io/projected/06dff8f3-1b1f-4905-8976-ad14fd8b0b3f-kube-api-access-xmvs9\") pod \"ovn-operator-controller-manager-b6456fdb6-zws5c\" (UID: \"06dff8f3-1b1f-4905-8976-ad14fd8b0b3f\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.162081 4946 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.162124 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert podName:ee6d6741-bc1e-4ed2-a54c-608c37d5da97 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:14.66211248 +0000 UTC m=+1147.458802589 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert") pod "openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" (UID: "ee6d6741-bc1e-4ed2-a54c-608c37d5da97") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.160085 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-mc9qv" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.177055 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.181949 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ff7f\" (UniqueName: \"kubernetes.io/projected/4bbd7e74-3c10-47ea-ac92-2845c974a128-kube-api-access-5ff7f\") pod \"placement-operator-controller-manager-78f8948974-p28rk\" (UID: \"4bbd7e74-3c10-47ea-ac92-2845c974a128\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.192753 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncqlp\" (UniqueName: \"kubernetes.io/projected/59735e89-ef8b-4887-9af2-d2bd5d3f7bb0-kube-api-access-ncqlp\") pod \"nova-operator-controller-manager-697bc559fc-g6cf2\" (UID: \"59735e89-ef8b-4887-9af2-d2bd5d3f7bb0\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.201545 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.212101 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm99k\" (UniqueName: \"kubernetes.io/projected/77e8b06b-4c6d-478a-ac26-230010446011-kube-api-access-hm99k\") pod \"octavia-operator-controller-manager-998648c74-xlsd9\" (UID: \"77e8b06b-4c6d-478a-ac26-230010446011\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.217282 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrrh6\" (UniqueName: \"kubernetes.io/projected/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-kube-api-access-zrrh6\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.217931 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhlqm\" (UniqueName: \"kubernetes.io/projected/ab0a2041-fd3d-49dc-b184-fdd603df844c-kube-api-access-mhlqm\") pod \"swift-operator-controller-manager-5f8c65bbfc-w9prn\" (UID: \"ab0a2041-fd3d-49dc-b184-fdd603df844c\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.218385 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmvs9\" (UniqueName: \"kubernetes.io/projected/06dff8f3-1b1f-4905-8976-ad14fd8b0b3f-kube-api-access-xmvs9\") pod \"ovn-operator-controller-manager-b6456fdb6-zws5c\" (UID: \"06dff8f3-1b1f-4905-8976-ad14fd8b0b3f\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.222281 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf6df\" (UniqueName: \"kubernetes.io/projected/f6febc9c-4437-4f94-8996-f4ece33a42f2-kube-api-access-vf6df\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-9n79w\" (UID: \"f6febc9c-4437-4f94-8996-f4ece33a42f2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.246719 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.248620 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.271855 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4g9p\" (UniqueName: \"kubernetes.io/projected/12f0ad36-d30d-4d23-8677-8d452c5ea978-kube-api-access-l4g9p\") pod \"test-operator-controller-manager-5854674fcc-l4hnh\" (UID: \"12f0ad36-d30d-4d23-8677-8d452c5ea978\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.271950 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lczbl\" (UniqueName: \"kubernetes.io/projected/a063e091-3cc0-4966-a2d7-69bcba9885f3-kube-api-access-lczbl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nq7sd\" (UID: \"a063e091-3cc0-4966-a2d7-69bcba9885f3\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.272998 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.298365 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.299380 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.308475 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.315571 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-sm54l" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.325104 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.346370 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.354369 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.373574 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.374695 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.376414 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4g9p\" (UniqueName: \"kubernetes.io/projected/12f0ad36-d30d-4d23-8677-8d452c5ea978-kube-api-access-l4g9p\") pod \"test-operator-controller-manager-5854674fcc-l4hnh\" (UID: \"12f0ad36-d30d-4d23-8677-8d452c5ea978\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.376463 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lczbl\" (UniqueName: \"kubernetes.io/projected/a063e091-3cc0-4966-a2d7-69bcba9885f3-kube-api-access-lczbl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nq7sd\" (UID: \"a063e091-3cc0-4966-a2d7-69bcba9885f3\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.376545 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.376596 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsjsm\" (UniqueName: \"kubernetes.io/projected/3efd94f9-2bf7-404c-ae24-6884ead0de15-kube-api-access-hsjsm\") pod \"watcher-operator-controller-manager-769dc69bc-5p4zw\" (UID: \"3efd94f9-2bf7-404c-ae24-6884ead0de15\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.380240 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.380445 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.380841 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-82jgj" Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.380870 4946 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.380918 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert podName:26f08bfc-aab0-4d18-a223-8e93309c15b0 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:15.38090304 +0000 UTC m=+1148.177593139 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert") pod "infra-operator-controller-manager-57548d458d-lvk95" (UID: "26f08bfc-aab0-4d18-a223-8e93309c15b0") : secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.392498 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.409207 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.410623 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4g9p\" (UniqueName: \"kubernetes.io/projected/12f0ad36-d30d-4d23-8677-8d452c5ea978-kube-api-access-l4g9p\") pod \"test-operator-controller-manager-5854674fcc-l4hnh\" (UID: \"12f0ad36-d30d-4d23-8677-8d452c5ea978\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.418373 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lczbl\" (UniqueName: \"kubernetes.io/projected/a063e091-3cc0-4966-a2d7-69bcba9885f3-kube-api-access-lczbl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nq7sd\" (UID: \"a063e091-3cc0-4966-a2d7-69bcba9885f3\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.442776 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.452690 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.453588 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.458698 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-6jjpv" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.464082 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.477348 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.477419 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqlvw\" (UniqueName: \"kubernetes.io/projected/7c469591-852f-4294-bd9e-8d1dc58e5d75-kube-api-access-wqlvw\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.477519 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsjsm\" (UniqueName: \"kubernetes.io/projected/3efd94f9-2bf7-404c-ae24-6884ead0de15-kube-api-access-hsjsm\") pod \"watcher-operator-controller-manager-769dc69bc-5p4zw\" (UID: \"3efd94f9-2bf7-404c-ae24-6884ead0de15\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.477565 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.485191 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.498428 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsjsm\" (UniqueName: \"kubernetes.io/projected/3efd94f9-2bf7-404c-ae24-6884ead0de15-kube-api-access-hsjsm\") pod \"watcher-operator-controller-manager-769dc69bc-5p4zw\" (UID: \"3efd94f9-2bf7-404c-ae24-6884ead0de15\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.511259 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.517664 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.578926 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqlvw\" (UniqueName: \"kubernetes.io/projected/7c469591-852f-4294-bd9e-8d1dc58e5d75-kube-api-access-wqlvw\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.579068 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.579112 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx2bh\" (UniqueName: \"kubernetes.io/projected/e20ea5bc-56c0-49ac-a27c-e6e93778a09a-kube-api-access-nx2bh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-z2k4w\" (UID: \"e20ea5bc-56c0-49ac-a27c-e6e93778a09a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.579134 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.579280 4946 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.579326 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:15.079312256 +0000 UTC m=+1147.876002365 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "webhook-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.579783 4946 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.579809 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:15.079801959 +0000 UTC m=+1147.876492068 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "metrics-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.614696 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.617149 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqlvw\" (UniqueName: \"kubernetes.io/projected/7c469591-852f-4294-bd9e-8d1dc58e5d75-kube-api-access-wqlvw\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.684186 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.684276 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx2bh\" (UniqueName: \"kubernetes.io/projected/e20ea5bc-56c0-49ac-a27c-e6e93778a09a-kube-api-access-nx2bh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-z2k4w\" (UID: \"e20ea5bc-56c0-49ac-a27c-e6e93778a09a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.684314 4946 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: E1203 07:09:14.684368 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert podName:ee6d6741-bc1e-4ed2-a54c-608c37d5da97 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:15.68435386 +0000 UTC m=+1148.481043969 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert") pod "openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" (UID: "ee6d6741-bc1e-4ed2-a54c-608c37d5da97") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.708810 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.714700 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m"] Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.716306 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx2bh\" (UniqueName: \"kubernetes.io/projected/e20ea5bc-56c0-49ac-a27c-e6e93778a09a-kube-api-access-nx2bh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-z2k4w\" (UID: \"e20ea5bc-56c0-49ac-a27c-e6e93778a09a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.771607 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.887667 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9"] Dec 03 07:09:14 crc kubenswrapper[4946]: W1203 07:09:14.919402 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafb5f2e7_fb0a_425b_9a3c_cff10fdec757.slice/crio-d533c05fb480f181e7365cd45c6afe964bc6188df41d88f0a92a53e8292da34c WatchSource:0}: Error finding container d533c05fb480f181e7365cd45c6afe964bc6188df41d88f0a92a53e8292da34c: Status 404 returned error can't find the container with id d533c05fb480f181e7365cd45c6afe964bc6188df41d88f0a92a53e8292da34c Dec 03 07:09:14 crc kubenswrapper[4946]: I1203 07:09:14.982543 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" event={"ID":"881fe69d-7e12-4250-8589-76096d3318d9","Type":"ContainerStarted","Data":"5a90ee6fc825884c82103660e298afa10c1f8b57da66d4f1b44615106c77d40b"} Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.000730 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.002441 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" event={"ID":"592abc1b-947b-4b47-965f-0110fa5e74e3","Type":"ContainerStarted","Data":"252c9099475061a157d889410153973c6ee5e9dbcf457c2ebbfcf44b61c45358"} Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.038414 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" event={"ID":"afb5f2e7-fb0a-425b-9a3c-cff10fdec757","Type":"ContainerStarted","Data":"d533c05fb480f181e7365cd45c6afe964bc6188df41d88f0a92a53e8292da34c"} Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.042344 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" event={"ID":"feaece87-4f9e-4945-8faa-c49487a57c2f","Type":"ContainerStarted","Data":"9dc6c38b44630df121cb3e484cee6bf382f855c27f5cb6ae7b4c867012b8035c"} Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.046260 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" event={"ID":"f050d6c8-74b6-42cb-943b-6e372c3e0dfe","Type":"ContainerStarted","Data":"174c5b6ff0c19675903874f3ca7fbb115838c1c00f41ed8b8eb8adb047cbd564"} Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.102521 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.102577 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.102753 4946 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.102817 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:16.10279751 +0000 UTC m=+1148.899487619 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "webhook-server-cert" not found Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.102973 4946 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.103047 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:16.103029236 +0000 UTC m=+1148.899719345 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "metrics-server-cert" not found Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.247790 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.255337 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.345905 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.352541 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.365752 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.371835 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.377187 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.407299 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.407510 4946 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.407576 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert podName:26f08bfc-aab0-4d18-a223-8e93309c15b0 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:17.407559345 +0000 UTC m=+1150.204249454 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert") pod "infra-operator-controller-manager-57548d458d-lvk95" (UID: "26f08bfc-aab0-4d18-a223-8e93309c15b0") : secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.586551 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.590711 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w"] Dec 03 07:09:15 crc kubenswrapper[4946]: W1203 07:09:15.605281 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77e8b06b_4c6d_478a_ac26_230010446011.slice/crio-671e5f9468d677002dfdcc95be8d6f5bbd1383c4765107020e6bfde90ef3c69a WatchSource:0}: Error finding container 671e5f9468d677002dfdcc95be8d6f5bbd1383c4765107020e6bfde90ef3c69a: Status 404 returned error can't find the container with id 671e5f9468d677002dfdcc95be8d6f5bbd1383c4765107020e6bfde90ef3c69a Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.608082 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn"] Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.610758 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nx2bh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-z2k4w_openstack-operators(e20ea5bc-56c0-49ac-a27c-e6e93778a09a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.612117 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" podUID="e20ea5bc-56c0-49ac-a27c-e6e93778a09a" Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.616292 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-p28rk"] Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.623001 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mhlqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-w9prn_openstack-operators(ab0a2041-fd3d-49dc-b184-fdd603df844c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.623232 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vf6df,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-9n79w_openstack-operators(f6febc9c-4437-4f94-8996-f4ece33a42f2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.637223 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw"] Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.643977 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hm99k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-xlsd9_openstack-operators(77e8b06b-4c6d-478a-ac26-230010446011): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.644198 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vf6df,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-9n79w_openstack-operators(f6febc9c-4437-4f94-8996-f4ece33a42f2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.644376 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mhlqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-w9prn_openstack-operators(ab0a2041-fd3d-49dc-b184-fdd603df844c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.646479 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hsjsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-5p4zw_openstack-operators(3efd94f9-2bf7-404c-ae24-6884ead0de15): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.646578 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" podUID="ab0a2041-fd3d-49dc-b184-fdd603df844c" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.646940 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" podUID="f6febc9c-4437-4f94-8996-f4ece33a42f2" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.679614 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hsjsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-5p4zw_openstack-operators(3efd94f9-2bf7-404c-ae24-6884ead0de15): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.679688 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hm99k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-xlsd9_openstack-operators(77e8b06b-4c6d-478a-ac26-230010446011): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.681057 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" podUID="77e8b06b-4c6d-478a-ac26-230010446011" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.681205 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" podUID="3efd94f9-2bf7-404c-ae24-6884ead0de15" Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.691050 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.710931 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9"] Dec 03 07:09:15 crc kubenswrapper[4946]: I1203 07:09:15.750376 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.750957 4946 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:15 crc kubenswrapper[4946]: E1203 07:09:15.751047 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert podName:ee6d6741-bc1e-4ed2-a54c-608c37d5da97 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:17.751024303 +0000 UTC m=+1150.547714442 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert") pod "openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" (UID: "ee6d6741-bc1e-4ed2-a54c-608c37d5da97") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.053036 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" event={"ID":"4110947d-509f-4bb7-bb92-e52ebb362340","Type":"ContainerStarted","Data":"0f7ef614822399f198f4583e4fbeb2d8fd0c6bf3643780f7b3bb73b7f97fb129"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.054559 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" event={"ID":"bd12d71a-063b-445c-876b-aab60ef74f35","Type":"ContainerStarted","Data":"6b56c2b42a1f08edfa572992490131212a21d62ccbea48615ab3f69e370e9cc7"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.056105 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" event={"ID":"4bbd7e74-3c10-47ea-ac92-2845c974a128","Type":"ContainerStarted","Data":"39f335e876949df9d950cefccb992f40de0e0cedf97964922c548333396fa476"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.060335 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" event={"ID":"a063e091-3cc0-4966-a2d7-69bcba9885f3","Type":"ContainerStarted","Data":"8ba7c5b77ecff1b8620e3277e81e97f697df53a8b052cd4e68918460b080d68a"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.062504 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" event={"ID":"9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81","Type":"ContainerStarted","Data":"ba629b6a5cf8a25e3b29a7fe9992b653572dc56fd65c4e0ff474d46f1d87877f"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.063514 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" event={"ID":"12f0ad36-d30d-4d23-8677-8d452c5ea978","Type":"ContainerStarted","Data":"cd2feefd3ea6edd83ee73c4bf35935952c76b0a9f3046a39d64988072305488b"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.065292 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" event={"ID":"06dff8f3-1b1f-4905-8976-ad14fd8b0b3f","Type":"ContainerStarted","Data":"8bb0b6011681cfe33cf395fa5781f3b7de299441ea0f4fbc7a639104cf01651f"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.072256 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" event={"ID":"29fe18fb-39d7-4003-8171-0e47aa1d1b18","Type":"ContainerStarted","Data":"887d8320c03330a67a91698b063888c3af8af933cb8ebb86e2d05cc4471d1a0f"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.073823 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" event={"ID":"59735e89-ef8b-4887-9af2-d2bd5d3f7bb0","Type":"ContainerStarted","Data":"aceca9e5701774dc657905c77daa129989771917c29ba4fbb4a2ba1a613335c3"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.075783 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" event={"ID":"f6febc9c-4437-4f94-8996-f4ece33a42f2","Type":"ContainerStarted","Data":"28b65ea229f5807192cbfddb79a3663c6375dd96779435dfa5010bb96a9e0d6c"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.082174 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" event={"ID":"f55d3e9f-d953-4672-9438-4f8db6a0af1b","Type":"ContainerStarted","Data":"3c241b68b275ad38a89ff0a8896d578a0d550050b35f1c263d9ccc905ea21cba"} Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.083064 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" podUID="f6febc9c-4437-4f94-8996-f4ece33a42f2" Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.084010 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" event={"ID":"e20ea5bc-56c0-49ac-a27c-e6e93778a09a","Type":"ContainerStarted","Data":"2a246ce37f3fa4b0a6d865ab5f5e2b4d5517b395bbd7c22711f1492cd49060be"} Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.085382 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" podUID="e20ea5bc-56c0-49ac-a27c-e6e93778a09a" Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.088330 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" event={"ID":"ab0a2041-fd3d-49dc-b184-fdd603df844c","Type":"ContainerStarted","Data":"81b1965581f89dbaed7e77d47e8fc04dd6ad1b96a35ed897cfa3d6f64d6aa9ef"} Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.091797 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" event={"ID":"3efd94f9-2bf7-404c-ae24-6884ead0de15","Type":"ContainerStarted","Data":"4ea915e9847f9bd533cca4050d6ade42ee205aa8605d19fa56e335a0a2944f79"} Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.091846 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" podUID="ab0a2041-fd3d-49dc-b184-fdd603df844c" Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.094379 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" event={"ID":"77e8b06b-4c6d-478a-ac26-230010446011","Type":"ContainerStarted","Data":"671e5f9468d677002dfdcc95be8d6f5bbd1383c4765107020e6bfde90ef3c69a"} Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.095188 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" podUID="3efd94f9-2bf7-404c-ae24-6884ead0de15" Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.097995 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" podUID="77e8b06b-4c6d-478a-ac26-230010446011" Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.154682 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:16 crc kubenswrapper[4946]: I1203 07:09:16.154812 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.154998 4946 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.154996 4946 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.155048 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:18.155029878 +0000 UTC m=+1150.951719987 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "webhook-server-cert" not found Dec 03 07:09:16 crc kubenswrapper[4946]: E1203 07:09:16.155232 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:18.155209073 +0000 UTC m=+1150.951899272 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "metrics-server-cert" not found Dec 03 07:09:17 crc kubenswrapper[4946]: E1203 07:09:17.110022 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" podUID="e20ea5bc-56c0-49ac-a27c-e6e93778a09a" Dec 03 07:09:17 crc kubenswrapper[4946]: E1203 07:09:17.110465 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" podUID="77e8b06b-4c6d-478a-ac26-230010446011" Dec 03 07:09:17 crc kubenswrapper[4946]: E1203 07:09:17.110505 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" podUID="3efd94f9-2bf7-404c-ae24-6884ead0de15" Dec 03 07:09:17 crc kubenswrapper[4946]: E1203 07:09:17.110852 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" podUID="ab0a2041-fd3d-49dc-b184-fdd603df844c" Dec 03 07:09:17 crc kubenswrapper[4946]: E1203 07:09:17.110895 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" podUID="f6febc9c-4437-4f94-8996-f4ece33a42f2" Dec 03 07:09:17 crc kubenswrapper[4946]: I1203 07:09:17.417035 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:17 crc kubenswrapper[4946]: E1203 07:09:17.417196 4946 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:17 crc kubenswrapper[4946]: E1203 07:09:17.417258 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert podName:26f08bfc-aab0-4d18-a223-8e93309c15b0 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:21.4172424 +0000 UTC m=+1154.213932509 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert") pod "infra-operator-controller-manager-57548d458d-lvk95" (UID: "26f08bfc-aab0-4d18-a223-8e93309c15b0") : secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:18 crc kubenswrapper[4946]: I1203 07:09:18.048890 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:18 crc kubenswrapper[4946]: E1203 07:09:18.049035 4946 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:18 crc kubenswrapper[4946]: E1203 07:09:18.049081 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert podName:ee6d6741-bc1e-4ed2-a54c-608c37d5da97 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:22.049067115 +0000 UTC m=+1154.845757224 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert") pod "openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" (UID: "ee6d6741-bc1e-4ed2-a54c-608c37d5da97") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:18 crc kubenswrapper[4946]: I1203 07:09:18.251512 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:18 crc kubenswrapper[4946]: I1203 07:09:18.251584 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:18 crc kubenswrapper[4946]: E1203 07:09:18.251753 4946 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 07:09:18 crc kubenswrapper[4946]: E1203 07:09:18.251821 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:22.251802737 +0000 UTC m=+1155.048492846 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "webhook-server-cert" not found Dec 03 07:09:18 crc kubenswrapper[4946]: E1203 07:09:18.251982 4946 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 07:09:18 crc kubenswrapper[4946]: E1203 07:09:18.252100 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:22.252057214 +0000 UTC m=+1155.048747393 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "metrics-server-cert" not found Dec 03 07:09:21 crc kubenswrapper[4946]: I1203 07:09:21.502173 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:21 crc kubenswrapper[4946]: E1203 07:09:21.502432 4946 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:21 crc kubenswrapper[4946]: E1203 07:09:21.502548 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert podName:26f08bfc-aab0-4d18-a223-8e93309c15b0 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:29.502520149 +0000 UTC m=+1162.299210338 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert") pod "infra-operator-controller-manager-57548d458d-lvk95" (UID: "26f08bfc-aab0-4d18-a223-8e93309c15b0") : secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:22 crc kubenswrapper[4946]: I1203 07:09:22.109011 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:22 crc kubenswrapper[4946]: E1203 07:09:22.109195 4946 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:22 crc kubenswrapper[4946]: E1203 07:09:22.109519 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert podName:ee6d6741-bc1e-4ed2-a54c-608c37d5da97 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:30.109497271 +0000 UTC m=+1162.906187380 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert") pod "openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" (UID: "ee6d6741-bc1e-4ed2-a54c-608c37d5da97") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:22 crc kubenswrapper[4946]: I1203 07:09:22.311235 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:22 crc kubenswrapper[4946]: I1203 07:09:22.311304 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:22 crc kubenswrapper[4946]: E1203 07:09:22.311521 4946 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 07:09:22 crc kubenswrapper[4946]: E1203 07:09:22.311582 4946 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 07:09:22 crc kubenswrapper[4946]: E1203 07:09:22.311590 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:30.311571155 +0000 UTC m=+1163.108261264 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "webhook-server-cert" not found Dec 03 07:09:22 crc kubenswrapper[4946]: E1203 07:09:22.311636 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs podName:7c469591-852f-4294-bd9e-8d1dc58e5d75 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:30.311624776 +0000 UTC m=+1163.108314885 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs") pod "openstack-operator-controller-manager-9f56fc979-wtn4r" (UID: "7c469591-852f-4294-bd9e-8d1dc58e5d75") : secret "metrics-server-cert" not found Dec 03 07:09:23 crc kubenswrapper[4946]: I1203 07:09:23.039293 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:09:23 crc kubenswrapper[4946]: I1203 07:09:23.039369 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:09:23 crc kubenswrapper[4946]: I1203 07:09:23.039424 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:09:23 crc kubenswrapper[4946]: I1203 07:09:23.040093 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5f3af879642bc444185e03b298fe438ec767c4e61a69937761fbbcf8110a3b1"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:09:23 crc kubenswrapper[4946]: I1203 07:09:23.040185 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://b5f3af879642bc444185e03b298fe438ec767c4e61a69937761fbbcf8110a3b1" gracePeriod=600 Dec 03 07:09:24 crc kubenswrapper[4946]: I1203 07:09:24.176552 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="b5f3af879642bc444185e03b298fe438ec767c4e61a69937761fbbcf8110a3b1" exitCode=0 Dec 03 07:09:24 crc kubenswrapper[4946]: I1203 07:09:24.176717 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"b5f3af879642bc444185e03b298fe438ec767c4e61a69937761fbbcf8110a3b1"} Dec 03 07:09:24 crc kubenswrapper[4946]: I1203 07:09:24.177274 4946 scope.go:117] "RemoveContainer" containerID="eb200c920ce9a9189cb9d22399d35826a7f0ee55cd4d5b661511822e8281f026" Dec 03 07:09:29 crc kubenswrapper[4946]: I1203 07:09:29.571504 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:29 crc kubenswrapper[4946]: E1203 07:09:29.571647 4946 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:29 crc kubenswrapper[4946]: E1203 07:09:29.572176 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert podName:26f08bfc-aab0-4d18-a223-8e93309c15b0 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:45.572161273 +0000 UTC m=+1178.368851392 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert") pod "infra-operator-controller-manager-57548d458d-lvk95" (UID: "26f08bfc-aab0-4d18-a223-8e93309c15b0") : secret "infra-operator-webhook-server-cert" not found Dec 03 07:09:30 crc kubenswrapper[4946]: E1203 07:09:30.160059 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 03 07:09:30 crc kubenswrapper[4946]: E1203 07:09:30.160288 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xmvs9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-zws5c_openstack-operators(06dff8f3-1b1f-4905-8976-ad14fd8b0b3f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:09:30 crc kubenswrapper[4946]: I1203 07:09:30.179115 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:30 crc kubenswrapper[4946]: E1203 07:09:30.180213 4946 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:30 crc kubenswrapper[4946]: E1203 07:09:30.180402 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert podName:ee6d6741-bc1e-4ed2-a54c-608c37d5da97 nodeName:}" failed. No retries permitted until 2025-12-03 07:09:46.180363028 +0000 UTC m=+1178.977053177 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert") pod "openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" (UID: "ee6d6741-bc1e-4ed2-a54c-608c37d5da97") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 07:09:30 crc kubenswrapper[4946]: I1203 07:09:30.382169 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:30 crc kubenswrapper[4946]: I1203 07:09:30.382232 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:30 crc kubenswrapper[4946]: I1203 07:09:30.388808 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-webhook-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:30 crc kubenswrapper[4946]: I1203 07:09:30.389229 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7c469591-852f-4294-bd9e-8d1dc58e5d75-metrics-certs\") pod \"openstack-operator-controller-manager-9f56fc979-wtn4r\" (UID: \"7c469591-852f-4294-bd9e-8d1dc58e5d75\") " pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:30 crc kubenswrapper[4946]: I1203 07:09:30.627032 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:30 crc kubenswrapper[4946]: E1203 07:09:30.835232 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 03 07:09:30 crc kubenswrapper[4946]: E1203 07:09:30.835823 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wwgrx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-rkv2m_openstack-operators(f050d6c8-74b6-42cb-943b-6e372c3e0dfe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:09:31 crc kubenswrapper[4946]: E1203 07:09:31.615793 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 03 07:09:31 crc kubenswrapper[4946]: E1203 07:09:31.615986 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k4cj6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-7cwn9_openstack-operators(afb5f2e7-fb0a-425b-9a3c-cff10fdec757): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:09:32 crc kubenswrapper[4946]: E1203 07:09:32.495959 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 07:09:32 crc kubenswrapper[4946]: E1203 07:09:32.496132 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ncqlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-g6cf2_openstack-operators(59735e89-ef8b-4887-9af2-d2bd5d3f7bb0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:09:33 crc kubenswrapper[4946]: E1203 07:09:33.088722 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 07:09:33 crc kubenswrapper[4946]: E1203 07:09:33.089266 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fn9qf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-b68fp_openstack-operators(f55d3e9f-d953-4672-9438-4f8db6a0af1b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:09:44 crc kubenswrapper[4946]: E1203 07:09:44.157539 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 03 07:09:44 crc kubenswrapper[4946]: E1203 07:09:44.158248 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mhlqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-w9prn_openstack-operators(ab0a2041-fd3d-49dc-b184-fdd603df844c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:09:44 crc kubenswrapper[4946]: E1203 07:09:44.600946 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 07:09:44 crc kubenswrapper[4946]: E1203 07:09:44.601312 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nx2bh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-z2k4w_openstack-operators(e20ea5bc-56c0-49ac-a27c-e6e93778a09a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:09:44 crc kubenswrapper[4946]: E1203 07:09:44.602452 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" podUID="e20ea5bc-56c0-49ac-a27c-e6e93778a09a" Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.225124 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r"] Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.394785 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" event={"ID":"4bbd7e74-3c10-47ea-ac92-2845c974a128","Type":"ContainerStarted","Data":"a8bb7af0cda1fa8ed8c6e8844ef26ecbba0ca4c41825c4ef1041882835f4b3d0"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.423906 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"a645c07ea34ddd66718ac1597141e393ef581a3ac2975f8c8d11ca9aa11d50ab"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.438164 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" event={"ID":"12f0ad36-d30d-4d23-8677-8d452c5ea978","Type":"ContainerStarted","Data":"27ad43e2768b2b9094df2466c53d9fdbce74496ec0fd72553ea5a604faf8b180"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.444167 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" event={"ID":"29fe18fb-39d7-4003-8171-0e47aa1d1b18","Type":"ContainerStarted","Data":"34f9049378900bbd66b78f4ec59930114d800f23128b15ed81e4a67143eab679"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.461698 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" event={"ID":"9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81","Type":"ContainerStarted","Data":"c37329145db886d964d59edff325647df5c0414e42a55c38b16f68d3f9d9f98f"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.481753 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" event={"ID":"4110947d-509f-4bb7-bb92-e52ebb362340","Type":"ContainerStarted","Data":"3f167bb8155ee85673a198e221ec86a5cbea34c1729051ca0743288492896536"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.504949 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" event={"ID":"bd12d71a-063b-445c-876b-aab60ef74f35","Type":"ContainerStarted","Data":"629973a357c710a0490f50200c87d26f4b8506bb565c263c99815189288bece8"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.544801 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" event={"ID":"a063e091-3cc0-4966-a2d7-69bcba9885f3","Type":"ContainerStarted","Data":"805707dcefbd5a04636817cfb332e42e4a3a505671da83ba351909f62ef679a7"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.564969 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" event={"ID":"feaece87-4f9e-4945-8faa-c49487a57c2f","Type":"ContainerStarted","Data":"5f8163280824fca5ae277f0b3ec009dfe3ad05fbea179ac02320dfa038020690"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.579024 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" event={"ID":"592abc1b-947b-4b47-965f-0110fa5e74e3","Type":"ContainerStarted","Data":"d7ccff4877a015a7f32c9a25af1e70e289d973f60e45d87cd1e19cb0b9b45f83"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.611171 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" event={"ID":"881fe69d-7e12-4250-8589-76096d3318d9","Type":"ContainerStarted","Data":"ef46026c6e7962d32496e3c3d982f9530a986d17f7fedc559589805c57e7e46a"} Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.643555 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.649637 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26f08bfc-aab0-4d18-a223-8e93309c15b0-cert\") pod \"infra-operator-controller-manager-57548d458d-lvk95\" (UID: \"26f08bfc-aab0-4d18-a223-8e93309c15b0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:45 crc kubenswrapper[4946]: I1203 07:09:45.817662 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:46 crc kubenswrapper[4946]: I1203 07:09:46.254843 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:46 crc kubenswrapper[4946]: I1203 07:09:46.284512 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee6d6741-bc1e-4ed2-a54c-608c37d5da97-cert\") pod \"openstack-baremetal-operator-controller-manager-55d86b6686k9mhl\" (UID: \"ee6d6741-bc1e-4ed2-a54c-608c37d5da97\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:46 crc kubenswrapper[4946]: I1203 07:09:46.428994 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:46 crc kubenswrapper[4946]: I1203 07:09:46.611844 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" event={"ID":"f6febc9c-4437-4f94-8996-f4ece33a42f2","Type":"ContainerStarted","Data":"8b8b49ee424f82de215671ba96d931a82da0fceeab1643817563d1a31f6226a3"} Dec 03 07:09:46 crc kubenswrapper[4946]: I1203 07:09:46.613711 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" event={"ID":"7c469591-852f-4294-bd9e-8d1dc58e5d75","Type":"ContainerStarted","Data":"a36adce34fd69665947c1750c5017debef2373f573646da0f8afe9e621cdb9c7"} Dec 03 07:09:46 crc kubenswrapper[4946]: I1203 07:09:46.618562 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" event={"ID":"3efd94f9-2bf7-404c-ae24-6884ead0de15","Type":"ContainerStarted","Data":"dfef0e4c1cf60b83f733cf619cc5a45beee4656221177d8ac3d495d7cd4df869"} Dec 03 07:09:47 crc kubenswrapper[4946]: I1203 07:09:47.356701 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-lvk95"] Dec 03 07:09:47 crc kubenswrapper[4946]: I1203 07:09:47.640634 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" event={"ID":"26f08bfc-aab0-4d18-a223-8e93309c15b0","Type":"ContainerStarted","Data":"cecf98d0533a85327c739c6cd7444615b8b9628167bf907c21ba08a26f54d5bf"} Dec 03 07:09:47 crc kubenswrapper[4946]: I1203 07:09:47.644476 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" event={"ID":"77e8b06b-4c6d-478a-ac26-230010446011","Type":"ContainerStarted","Data":"3f0722a70c6a241c9b42b310ed90230a0302230ee34e19397d7e15de904853f0"} Dec 03 07:09:47 crc kubenswrapper[4946]: I1203 07:09:47.646805 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" event={"ID":"7c469591-852f-4294-bd9e-8d1dc58e5d75","Type":"ContainerStarted","Data":"a19074847234307e0b1ae90f3ec5c65242766b938f6401a8dbf4b5f8a58c8655"} Dec 03 07:09:47 crc kubenswrapper[4946]: I1203 07:09:47.647118 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:09:47 crc kubenswrapper[4946]: I1203 07:09:47.735474 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" podStartSLOduration=33.7354548 podStartE2EDuration="33.7354548s" podCreationTimestamp="2025-12-03 07:09:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:09:47.725034572 +0000 UTC m=+1180.521724701" watchObservedRunningTime="2025-12-03 07:09:47.7354548 +0000 UTC m=+1180.532144919" Dec 03 07:09:47 crc kubenswrapper[4946]: I1203 07:09:47.794951 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl"] Dec 03 07:09:48 crc kubenswrapper[4946]: I1203 07:09:48.690049 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" event={"ID":"ee6d6741-bc1e-4ed2-a54c-608c37d5da97","Type":"ContainerStarted","Data":"7f8b6f590a82de3afd13eca9a6846f127225f70d97f31d49f652e33d5229fd9e"} Dec 03 07:09:50 crc kubenswrapper[4946]: E1203 07:09:50.747448 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" podUID="f050d6c8-74b6-42cb-943b-6e372c3e0dfe" Dec 03 07:09:50 crc kubenswrapper[4946]: E1203 07:09:50.898863 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" podUID="06dff8f3-1b1f-4905-8976-ad14fd8b0b3f" Dec 03 07:09:51 crc kubenswrapper[4946]: E1203 07:09:51.621430 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" podUID="ab0a2041-fd3d-49dc-b184-fdd603df844c" Dec 03 07:09:51 crc kubenswrapper[4946]: E1203 07:09:51.698073 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" podUID="afb5f2e7-fb0a-425b-9a3c-cff10fdec757" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.712301 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" event={"ID":"3efd94f9-2bf7-404c-ae24-6884ead0de15","Type":"ContainerStarted","Data":"e042b15daf999255172b77aa423cad54256ca920fa436f2b08c29c0b183480ca"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.713298 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.714729 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.721379 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" event={"ID":"f6febc9c-4437-4f94-8996-f4ece33a42f2","Type":"ContainerStarted","Data":"dac8f9eb29d31494d055effb6993d9cd08cd468c6e4bc76f5553f9e31cba0b87"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.722028 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.723336 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.724688 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" event={"ID":"06dff8f3-1b1f-4905-8976-ad14fd8b0b3f","Type":"ContainerStarted","Data":"4d4f8754db50bbf7d82e785081908f20175cebaca1674285c5cded95c9f5fa79"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.732729 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-5p4zw" podStartSLOduration=2.857654333 podStartE2EDuration="37.73271727s" podCreationTimestamp="2025-12-03 07:09:14 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.644169291 +0000 UTC m=+1148.440859400" lastFinishedPulling="2025-12-03 07:09:50.519232228 +0000 UTC m=+1183.315922337" observedRunningTime="2025-12-03 07:09:51.732015611 +0000 UTC m=+1184.528705720" watchObservedRunningTime="2025-12-03 07:09:51.73271727 +0000 UTC m=+1184.529407379" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.734136 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" event={"ID":"9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81","Type":"ContainerStarted","Data":"dd9451b9b3e9a13f4a89077b57280ccda7e1e1022a02ae99b86b414b39366925"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.735064 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.742195 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.742224 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" event={"ID":"bd12d71a-063b-445c-876b-aab60ef74f35","Type":"ContainerStarted","Data":"6c1273c8ed72fea1a103a8de7e562b2fa97b4eda214a5b5fe4e1c5b50bef0270"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.743426 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.743820 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.757247 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" event={"ID":"ab0a2041-fd3d-49dc-b184-fdd603df844c","Type":"ContainerStarted","Data":"457deb67a7dccaedd640c08cbad041c1459ea31ed9bd6d5638f458fe68bcbec9"} Dec 03 07:09:51 crc kubenswrapper[4946]: E1203 07:09:51.760172 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" podUID="ab0a2041-fd3d-49dc-b184-fdd603df844c" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.788699 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" event={"ID":"12f0ad36-d30d-4d23-8677-8d452c5ea978","Type":"ContainerStarted","Data":"7b3a897dce5bbee619c62e284b997cb9b39b7effd30b9a91d3fc1fe6469ede11"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.789351 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.791053 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-9n79w" podStartSLOduration=3.374750334 podStartE2EDuration="38.791028737s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.623102949 +0000 UTC m=+1148.419793058" lastFinishedPulling="2025-12-03 07:09:51.039381352 +0000 UTC m=+1183.836071461" observedRunningTime="2025-12-03 07:09:51.756102144 +0000 UTC m=+1184.552792253" watchObservedRunningTime="2025-12-03 07:09:51.791028737 +0000 UTC m=+1184.587718856" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.796005 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" event={"ID":"881fe69d-7e12-4250-8589-76096d3318d9","Type":"ContainerStarted","Data":"ca68f96beda5d3654f4c3b7054f5e3ad70cd5ac29cfe62a1546e1a3aeea4e932"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.796922 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.798920 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.799559 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" event={"ID":"a063e091-3cc0-4966-a2d7-69bcba9885f3","Type":"ContainerStarted","Data":"2a0985d1b975eda2cfec776ff1fc0b447704cbab29438411b1b11f3d8b3b9d66"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.800211 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.804038 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.815374 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.820960 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" event={"ID":"29fe18fb-39d7-4003-8171-0e47aa1d1b18","Type":"ContainerStarted","Data":"0fbf2ac672f2fe687a22ff42ca8f8a09b745556a9358adce439ce4d3e9096827"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.821804 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" Dec 03 07:09:51 crc kubenswrapper[4946]: E1203 07:09:51.826543 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" podUID="f55d3e9f-d953-4672-9438-4f8db6a0af1b" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.837970 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-sv727" podStartSLOduration=2.917403545 podStartE2EDuration="38.837955369s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.015993233 +0000 UTC m=+1147.812683342" lastFinishedPulling="2025-12-03 07:09:50.936545057 +0000 UTC m=+1183.733235166" observedRunningTime="2025-12-03 07:09:51.83762664 +0000 UTC m=+1184.634316749" watchObservedRunningTime="2025-12-03 07:09:51.837955369 +0000 UTC m=+1184.634645478" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.840250 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.841245 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" event={"ID":"4110947d-509f-4bb7-bb92-e52ebb362340","Type":"ContainerStarted","Data":"1a5258304a7680e4017843e06f962f64f897495c4ac04d02c40b951a94589177"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.841806 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.843553 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.852164 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" event={"ID":"afb5f2e7-fb0a-425b-9a3c-cff10fdec757","Type":"ContainerStarted","Data":"a97a1cbfdb45e6b11de217a569a4d6b554781ee864b5500f741947e6536d5492"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.864227 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" event={"ID":"f050d6c8-74b6-42cb-943b-6e372c3e0dfe","Type":"ContainerStarted","Data":"1ddf502ffd8c80f6c73d30e45d19cb505bc9d025d665091874a7d1a816b78c23"} Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.874672 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-v62hf" podStartSLOduration=2.254812278 podStartE2EDuration="38.874645979s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:14.657108503 +0000 UTC m=+1147.453798612" lastFinishedPulling="2025-12-03 07:09:51.276942204 +0000 UTC m=+1184.073632313" observedRunningTime="2025-12-03 07:09:51.87096981 +0000 UTC m=+1184.667659919" watchObservedRunningTime="2025-12-03 07:09:51.874645979 +0000 UTC m=+1184.671336088" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.926542 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-q27df" podStartSLOduration=2.964165183 podStartE2EDuration="38.926520013s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.377128133 +0000 UTC m=+1148.173818242" lastFinishedPulling="2025-12-03 07:09:51.339482963 +0000 UTC m=+1184.136173072" observedRunningTime="2025-12-03 07:09:51.901362282 +0000 UTC m=+1184.698052391" watchObservedRunningTime="2025-12-03 07:09:51.926520013 +0000 UTC m=+1184.723210122" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.958193 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nq7sd" podStartSLOduration=3.315874231 podStartE2EDuration="38.958169038s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.605294163 +0000 UTC m=+1148.401984272" lastFinishedPulling="2025-12-03 07:09:51.24758897 +0000 UTC m=+1184.044279079" observedRunningTime="2025-12-03 07:09:51.950247317 +0000 UTC m=+1184.746937426" watchObservedRunningTime="2025-12-03 07:09:51.958169038 +0000 UTC m=+1184.754859147" Dec 03 07:09:51 crc kubenswrapper[4946]: I1203 07:09:51.971976 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-l4hnh" podStartSLOduration=2.69908895 podStartE2EDuration="37.971962356s" podCreationTimestamp="2025-12-03 07:09:14 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.377544964 +0000 UTC m=+1148.174235073" lastFinishedPulling="2025-12-03 07:09:50.65041837 +0000 UTC m=+1183.447108479" observedRunningTime="2025-12-03 07:09:51.968896955 +0000 UTC m=+1184.765587064" watchObservedRunningTime="2025-12-03 07:09:51.971962356 +0000 UTC m=+1184.768652465" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.029962 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-q88nf" podStartSLOduration=3.46601594 podStartE2EDuration="39.029945504s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.273195139 +0000 UTC m=+1148.069885248" lastFinishedPulling="2025-12-03 07:09:50.837124703 +0000 UTC m=+1183.633814812" observedRunningTime="2025-12-03 07:09:52.029373669 +0000 UTC m=+1184.826063798" watchObservedRunningTime="2025-12-03 07:09:52.029945504 +0000 UTC m=+1184.826635613" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.102456 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2hnpw" podStartSLOduration=3.895988087 podStartE2EDuration="39.102436579s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.377460512 +0000 UTC m=+1148.174150621" lastFinishedPulling="2025-12-03 07:09:50.583909004 +0000 UTC m=+1183.380599113" observedRunningTime="2025-12-03 07:09:52.092628147 +0000 UTC m=+1184.889318256" watchObservedRunningTime="2025-12-03 07:09:52.102436579 +0000 UTC m=+1184.899126688" Dec 03 07:09:52 crc kubenswrapper[4946]: E1203 07:09:52.432370 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" podUID="59735e89-ef8b-4887-9af2-d2bd5d3f7bb0" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.878181 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" event={"ID":"f55d3e9f-d953-4672-9438-4f8db6a0af1b","Type":"ContainerStarted","Data":"80a9cd814d393ab82b68ef72a1c1715edc4c98370972918597013968cc7f917d"} Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.887866 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" event={"ID":"afb5f2e7-fb0a-425b-9a3c-cff10fdec757","Type":"ContainerStarted","Data":"23c00b8aa0ee319b6dabd775d4cdde267595655a410c6be6db659109fd39bcc7"} Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.888706 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.890726 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" event={"ID":"4bbd7e74-3c10-47ea-ac92-2845c974a128","Type":"ContainerStarted","Data":"5a7fe9567f387725a03095889b3156b796a39bc7fd52722f0312fee938600888"} Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.891345 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.892122 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" event={"ID":"77e8b06b-4c6d-478a-ac26-230010446011","Type":"ContainerStarted","Data":"31bec5ab7d8c0ec5833efdec807c4117b5c6897e413ce4cb674fc18a5ae5f13d"} Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.892649 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.893780 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" event={"ID":"06dff8f3-1b1f-4905-8976-ad14fd8b0b3f","Type":"ContainerStarted","Data":"e6afe48876d9996b7730cc26fb2fa36a51f45357295ad8eca8ce476ab4a2a4d6"} Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.893830 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.895298 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.899061 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" event={"ID":"feaece87-4f9e-4945-8faa-c49487a57c2f","Type":"ContainerStarted","Data":"9dcdb05a0e65a5e953893f264886f36d1b2cdf77392f7377a3125f96bc03dafe"} Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.900568 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.900781 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.903088 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.907342 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" event={"ID":"59735e89-ef8b-4887-9af2-d2bd5d3f7bb0","Type":"ContainerStarted","Data":"e40a1ccd0bf4f0e31b8812a454d6264bfd956ed18a69387f7580a0211ac101ea"} Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.947456 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" event={"ID":"592abc1b-947b-4b47-965f-0110fa5e74e3","Type":"ContainerStarted","Data":"fa2c3367660cb1c0b9399afe9e20ee3b82fd7616595b102c539b23ff34b68792"} Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.948435 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.948684 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" podStartSLOduration=2.495249166 podStartE2EDuration="39.948663328s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:14.929651258 +0000 UTC m=+1147.726341367" lastFinishedPulling="2025-12-03 07:09:52.38306542 +0000 UTC m=+1185.179755529" observedRunningTime="2025-12-03 07:09:52.946426328 +0000 UTC m=+1185.743116437" watchObservedRunningTime="2025-12-03 07:09:52.948663328 +0000 UTC m=+1185.745353447" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.950306 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" Dec 03 07:09:52 crc kubenswrapper[4946]: I1203 07:09:52.956093 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" event={"ID":"f050d6c8-74b6-42cb-943b-6e372c3e0dfe","Type":"ContainerStarted","Data":"6fe82f1542ceec76b4799834c9d4bfa1bc4767de569980d78b38f9c0418e22a1"} Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.025103 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-p28rk" podStartSLOduration=3.6450695189999998 podStartE2EDuration="40.025089388s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.609327521 +0000 UTC m=+1148.406017630" lastFinishedPulling="2025-12-03 07:09:51.98934739 +0000 UTC m=+1184.786037499" observedRunningTime="2025-12-03 07:09:53.023592508 +0000 UTC m=+1185.820282617" watchObservedRunningTime="2025-12-03 07:09:53.025089388 +0000 UTC m=+1185.821779497" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.025861 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-fwbh4" podStartSLOduration=3.021715029 podStartE2EDuration="40.025854428s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:14.837716874 +0000 UTC m=+1147.634406983" lastFinishedPulling="2025-12-03 07:09:51.841856273 +0000 UTC m=+1184.638546382" observedRunningTime="2025-12-03 07:09:52.989872348 +0000 UTC m=+1185.786562457" watchObservedRunningTime="2025-12-03 07:09:53.025854428 +0000 UTC m=+1185.822544537" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.067627 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" podStartSLOduration=3.181069153 podStartE2EDuration="40.067610653s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.379891697 +0000 UTC m=+1148.176581806" lastFinishedPulling="2025-12-03 07:09:52.266433197 +0000 UTC m=+1185.063123306" observedRunningTime="2025-12-03 07:09:53.066416371 +0000 UTC m=+1185.863106480" watchObservedRunningTime="2025-12-03 07:09:53.067610653 +0000 UTC m=+1185.864300762" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.085507 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-xlsd9" podStartSLOduration=3.566947863 podStartE2EDuration="40.08549014s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.643775481 +0000 UTC m=+1148.440465590" lastFinishedPulling="2025-12-03 07:09:52.162317758 +0000 UTC m=+1184.959007867" observedRunningTime="2025-12-03 07:09:53.082634764 +0000 UTC m=+1185.879324873" watchObservedRunningTime="2025-12-03 07:09:53.08549014 +0000 UTC m=+1185.882180249" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.116920 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-v7djg" podStartSLOduration=2.899100876 podStartE2EDuration="40.116900539s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:14.538216589 +0000 UTC m=+1147.334906698" lastFinishedPulling="2025-12-03 07:09:51.756016252 +0000 UTC m=+1184.552706361" observedRunningTime="2025-12-03 07:09:53.115242384 +0000 UTC m=+1185.911932503" watchObservedRunningTime="2025-12-03 07:09:53.116900539 +0000 UTC m=+1185.913590648" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.902255 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.962633 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" event={"ID":"59735e89-ef8b-4887-9af2-d2bd5d3f7bb0","Type":"ContainerStarted","Data":"8c339cec54217ea1cdb87525e921c050c1d558b7d24bd6a6137b1f4e92431564"} Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.963866 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.965114 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" event={"ID":"26f08bfc-aab0-4d18-a223-8e93309c15b0","Type":"ContainerStarted","Data":"f98a182f0e6bfa8fd964cbc821c2add657fac59a446cabefc318432a33feef17"} Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.966867 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" event={"ID":"f55d3e9f-d953-4672-9438-4f8db6a0af1b","Type":"ContainerStarted","Data":"04bc5e51c9378db63cf609110b4054f89e9829001bc6d9834739ae7b8ec7d213"} Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.966893 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.987167 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" podStartSLOduration=3.402360179 podStartE2EDuration="40.987145468s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:14.886749873 +0000 UTC m=+1147.683439982" lastFinishedPulling="2025-12-03 07:09:52.471535162 +0000 UTC m=+1185.268225271" observedRunningTime="2025-12-03 07:09:53.158419057 +0000 UTC m=+1185.955109166" watchObservedRunningTime="2025-12-03 07:09:53.987145468 +0000 UTC m=+1186.783835587" Dec 03 07:09:53 crc kubenswrapper[4946]: I1203 07:09:53.989789 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" podStartSLOduration=2.635166241 podStartE2EDuration="40.989770808s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.377258716 +0000 UTC m=+1148.173948825" lastFinishedPulling="2025-12-03 07:09:53.731863283 +0000 UTC m=+1186.528553392" observedRunningTime="2025-12-03 07:09:53.985917325 +0000 UTC m=+1186.782607454" watchObservedRunningTime="2025-12-03 07:09:53.989770808 +0000 UTC m=+1186.786460917" Dec 03 07:09:54 crc kubenswrapper[4946]: I1203 07:09:54.032229 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" podStartSLOduration=2.574875082 podStartE2EDuration="41.032216331s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.272184352 +0000 UTC m=+1148.068874461" lastFinishedPulling="2025-12-03 07:09:53.729525601 +0000 UTC m=+1186.526215710" observedRunningTime="2025-12-03 07:09:54.030561216 +0000 UTC m=+1186.827251325" watchObservedRunningTime="2025-12-03 07:09:54.032216331 +0000 UTC m=+1186.828906440" Dec 03 07:09:54 crc kubenswrapper[4946]: I1203 07:09:54.998591 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" event={"ID":"26f08bfc-aab0-4d18-a223-8e93309c15b0","Type":"ContainerStarted","Data":"591a35444bc7fee535a29eea1e6bc251db23db8ddea6e884cddf35570609dcdb"} Dec 03 07:09:54 crc kubenswrapper[4946]: I1203 07:09:54.998892 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:09:55 crc kubenswrapper[4946]: I1203 07:09:55.001208 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" event={"ID":"ee6d6741-bc1e-4ed2-a54c-608c37d5da97","Type":"ContainerStarted","Data":"5d5d03bcab6c6c21712974d2bea32674d8672de469d819217439a78632ba4d0d"} Dec 03 07:09:55 crc kubenswrapper[4946]: I1203 07:09:55.029242 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" podStartSLOduration=35.686176458 podStartE2EDuration="42.029226644s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:47.386141836 +0000 UTC m=+1180.182831935" lastFinishedPulling="2025-12-03 07:09:53.729192012 +0000 UTC m=+1186.525882121" observedRunningTime="2025-12-03 07:09:55.029042409 +0000 UTC m=+1187.825732518" watchObservedRunningTime="2025-12-03 07:09:55.029226644 +0000 UTC m=+1187.825916753" Dec 03 07:09:56 crc kubenswrapper[4946]: I1203 07:09:56.009125 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" event={"ID":"ee6d6741-bc1e-4ed2-a54c-608c37d5da97","Type":"ContainerStarted","Data":"68ad0d815ae7472680b6539dd13b4d3b6726143b400d83a8605e145f293fdf07"} Dec 03 07:09:56 crc kubenswrapper[4946]: I1203 07:09:56.430153 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:09:56 crc kubenswrapper[4946]: E1203 07:09:56.594792 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" podUID="e20ea5bc-56c0-49ac-a27c-e6e93778a09a" Dec 03 07:09:56 crc kubenswrapper[4946]: I1203 07:09:56.616285 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" podStartSLOduration=36.763452775 podStartE2EDuration="43.616261178s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:47.806221719 +0000 UTC m=+1180.602911828" lastFinishedPulling="2025-12-03 07:09:54.659030122 +0000 UTC m=+1187.455720231" observedRunningTime="2025-12-03 07:09:56.041649139 +0000 UTC m=+1188.838339258" watchObservedRunningTime="2025-12-03 07:09:56.616261178 +0000 UTC m=+1189.412951327" Dec 03 07:10:00 crc kubenswrapper[4946]: I1203 07:10:00.636632 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-9f56fc979-wtn4r" Dec 03 07:10:03 crc kubenswrapper[4946]: I1203 07:10:03.905286 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-rkv2m" Dec 03 07:10:03 crc kubenswrapper[4946]: I1203 07:10:03.939644 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7cwn9" Dec 03 07:10:04 crc kubenswrapper[4946]: I1203 07:10:04.145032 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-b68fp" Dec 03 07:10:04 crc kubenswrapper[4946]: I1203 07:10:04.305468 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g6cf2" Dec 03 07:10:04 crc kubenswrapper[4946]: I1203 07:10:04.351574 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-zws5c" Dec 03 07:10:05 crc kubenswrapper[4946]: I1203 07:10:05.829340 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lvk95" Dec 03 07:10:06 crc kubenswrapper[4946]: I1203 07:10:06.435824 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55d86b6686k9mhl" Dec 03 07:10:09 crc kubenswrapper[4946]: I1203 07:10:09.136915 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" event={"ID":"ab0a2041-fd3d-49dc-b184-fdd603df844c","Type":"ContainerStarted","Data":"aa9a92e4bfe344c62370f274402bb12d9129bb8d2101f5711570d6bcbfdf6ace"} Dec 03 07:10:09 crc kubenswrapper[4946]: I1203 07:10:09.137859 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" Dec 03 07:10:09 crc kubenswrapper[4946]: I1203 07:10:09.171365 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" podStartSLOduration=3.646769824 podStartE2EDuration="56.171340252s" podCreationTimestamp="2025-12-03 07:09:13 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.622845232 +0000 UTC m=+1148.419535341" lastFinishedPulling="2025-12-03 07:10:08.14741562 +0000 UTC m=+1200.944105769" observedRunningTime="2025-12-03 07:10:09.160576544 +0000 UTC m=+1201.957266733" watchObservedRunningTime="2025-12-03 07:10:09.171340252 +0000 UTC m=+1201.968030401" Dec 03 07:10:11 crc kubenswrapper[4946]: I1203 07:10:11.161657 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" event={"ID":"e20ea5bc-56c0-49ac-a27c-e6e93778a09a","Type":"ContainerStarted","Data":"a6192c3f7452f7f34c504efb35eff4c121faf89162f773736fa88935755f25aa"} Dec 03 07:10:11 crc kubenswrapper[4946]: I1203 07:10:11.191155 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-z2k4w" podStartSLOduration=2.766028198 podStartE2EDuration="57.191123427s" podCreationTimestamp="2025-12-03 07:09:14 +0000 UTC" firstStartedPulling="2025-12-03 07:09:15.610521823 +0000 UTC m=+1148.407211932" lastFinishedPulling="2025-12-03 07:10:10.035617012 +0000 UTC m=+1202.832307161" observedRunningTime="2025-12-03 07:10:11.182225239 +0000 UTC m=+1203.978915388" watchObservedRunningTime="2025-12-03 07:10:11.191123427 +0000 UTC m=+1203.987813576" Dec 03 07:10:14 crc kubenswrapper[4946]: I1203 07:10:14.413566 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-w9prn" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.516016 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-9jd5f"] Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.517966 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.521318 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.521508 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-kf484" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.521714 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.521810 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.538401 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-9jd5f"] Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.563691 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-567c455747-jh9sl"] Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.564816 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.566515 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.581979 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-jh9sl"] Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.615424 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdg4c\" (UniqueName: \"kubernetes.io/projected/c71804ce-5cc9-40cb-bc49-c1d5620cc034-kube-api-access-hdg4c\") pod \"dnsmasq-dns-5cd484bb89-9jd5f\" (UID: \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\") " pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.615469 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-config\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.615498 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71804ce-5cc9-40cb-bc49-c1d5620cc034-config\") pod \"dnsmasq-dns-5cd484bb89-9jd5f\" (UID: \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\") " pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.615559 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv2pm\" (UniqueName: \"kubernetes.io/projected/f22bc63b-9a11-4393-8976-15a2eb992ee2-kube-api-access-xv2pm\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.615578 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-dns-svc\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.716497 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71804ce-5cc9-40cb-bc49-c1d5620cc034-config\") pod \"dnsmasq-dns-5cd484bb89-9jd5f\" (UID: \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\") " pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.717267 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71804ce-5cc9-40cb-bc49-c1d5620cc034-config\") pod \"dnsmasq-dns-5cd484bb89-9jd5f\" (UID: \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\") " pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.717405 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv2pm\" (UniqueName: \"kubernetes.io/projected/f22bc63b-9a11-4393-8976-15a2eb992ee2-kube-api-access-xv2pm\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.717429 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-dns-svc\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.717836 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdg4c\" (UniqueName: \"kubernetes.io/projected/c71804ce-5cc9-40cb-bc49-c1d5620cc034-kube-api-access-hdg4c\") pod \"dnsmasq-dns-5cd484bb89-9jd5f\" (UID: \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\") " pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.718089 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-config\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.718317 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-dns-svc\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.718832 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-config\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.736770 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdg4c\" (UniqueName: \"kubernetes.io/projected/c71804ce-5cc9-40cb-bc49-c1d5620cc034-kube-api-access-hdg4c\") pod \"dnsmasq-dns-5cd484bb89-9jd5f\" (UID: \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\") " pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.738401 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv2pm\" (UniqueName: \"kubernetes.io/projected/f22bc63b-9a11-4393-8976-15a2eb992ee2-kube-api-access-xv2pm\") pod \"dnsmasq-dns-567c455747-jh9sl\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.892108 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:30 crc kubenswrapper[4946]: I1203 07:10:30.904726 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:31 crc kubenswrapper[4946]: I1203 07:10:31.392609 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-9jd5f"] Dec 03 07:10:31 crc kubenswrapper[4946]: W1203 07:10:31.396689 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc71804ce_5cc9_40cb_bc49_c1d5620cc034.slice/crio-77b296145fdf825fd5e7c7b3203b0a6efad7f85b71b0b4ce3ac565b07a68980f WatchSource:0}: Error finding container 77b296145fdf825fd5e7c7b3203b0a6efad7f85b71b0b4ce3ac565b07a68980f: Status 404 returned error can't find the container with id 77b296145fdf825fd5e7c7b3203b0a6efad7f85b71b0b4ce3ac565b07a68980f Dec 03 07:10:31 crc kubenswrapper[4946]: I1203 07:10:31.463814 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-jh9sl"] Dec 03 07:10:31 crc kubenswrapper[4946]: W1203 07:10:31.470555 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf22bc63b_9a11_4393_8976_15a2eb992ee2.slice/crio-bbb75635389a2b5728a1b3fa784ebbc1d3df794e85ca6893c358cd79bf279f6a WatchSource:0}: Error finding container bbb75635389a2b5728a1b3fa784ebbc1d3df794e85ca6893c358cd79bf279f6a: Status 404 returned error can't find the container with id bbb75635389a2b5728a1b3fa784ebbc1d3df794e85ca6893c358cd79bf279f6a Dec 03 07:10:32 crc kubenswrapper[4946]: I1203 07:10:32.363692 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-jh9sl" event={"ID":"f22bc63b-9a11-4393-8976-15a2eb992ee2","Type":"ContainerStarted","Data":"bbb75635389a2b5728a1b3fa784ebbc1d3df794e85ca6893c358cd79bf279f6a"} Dec 03 07:10:32 crc kubenswrapper[4946]: I1203 07:10:32.365238 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" event={"ID":"c71804ce-5cc9-40cb-bc49-c1d5620cc034","Type":"ContainerStarted","Data":"77b296145fdf825fd5e7c7b3203b0a6efad7f85b71b0b4ce3ac565b07a68980f"} Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.417274 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-jh9sl"] Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.447539 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-xnm28"] Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.448893 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.456372 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-xnm28"] Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.561718 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz7j2\" (UniqueName: \"kubernetes.io/projected/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-kube-api-access-fz7j2\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.561879 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-config\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.561986 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.663219 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-config\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.663279 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.663348 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz7j2\" (UniqueName: \"kubernetes.io/projected/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-kube-api-access-fz7j2\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.664324 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-config\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.664397 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.687896 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz7j2\" (UniqueName: \"kubernetes.io/projected/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-kube-api-access-fz7j2\") pod \"dnsmasq-dns-bc4b48fc9-xnm28\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.708752 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-9jd5f"] Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.754504 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb666b895-jjf2z"] Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.756762 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-jjf2z"] Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.756856 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.775112 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.865233 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-config\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.865532 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5j2t\" (UniqueName: \"kubernetes.io/projected/4ba0e22f-5c32-4638-8347-013fa968e479-kube-api-access-v5j2t\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.865582 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-dns-svc\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.966994 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-config\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.967033 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5j2t\" (UniqueName: \"kubernetes.io/projected/4ba0e22f-5c32-4638-8347-013fa968e479-kube-api-access-v5j2t\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.967085 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-dns-svc\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.967954 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-dns-svc\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.968441 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-config\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:33 crc kubenswrapper[4946]: I1203 07:10:33.988558 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5j2t\" (UniqueName: \"kubernetes.io/projected/4ba0e22f-5c32-4638-8347-013fa968e479-kube-api-access-v5j2t\") pod \"dnsmasq-dns-cb666b895-jjf2z\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.083625 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.257751 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-xnm28"] Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.395271 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" event={"ID":"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb","Type":"ContainerStarted","Data":"c97d21d0acf1eb760e9738395def00851124acb199f8c89252d63bc5856f609a"} Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.547964 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-jjf2z"] Dec 03 07:10:34 crc kubenswrapper[4946]: W1203 07:10:34.555183 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ba0e22f_5c32_4638_8347_013fa968e479.slice/crio-a5de8e034bcc3aed831766d620d3865eafa93abda2d9701456283decb0082c29 WatchSource:0}: Error finding container a5de8e034bcc3aed831766d620d3865eafa93abda2d9701456283decb0082c29: Status 404 returned error can't find the container with id a5de8e034bcc3aed831766d620d3865eafa93abda2d9701456283decb0082c29 Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.606397 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.607712 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.609422 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.610449 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.610450 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.610461 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-97lbv" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.611172 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.611321 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.616108 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.630708 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.775824 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.775883 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/854575dc-d77d-43fe-a7ff-495ef76ab456-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.775913 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.775939 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.775962 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/854575dc-d77d-43fe-a7ff-495ef76ab456-pod-info\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.777027 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqpbx\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-kube-api-access-rqpbx\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.777198 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.777251 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-server-conf\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.777282 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.777899 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.778017 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.881954 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882015 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882038 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882056 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882090 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/854575dc-d77d-43fe-a7ff-495ef76ab456-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882114 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882131 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882151 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/854575dc-d77d-43fe-a7ff-495ef76ab456-pod-info\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882171 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqpbx\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-kube-api-access-rqpbx\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882211 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882236 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-server-conf\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.882634 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.885477 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.885609 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.886356 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.886793 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-server-conf\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.887098 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.888414 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.890471 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.895252 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.895432 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.895595 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.895693 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.895903 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.896575 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g5m7z" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.896687 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.897837 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.903420 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.903863 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqpbx\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-kube-api-access-rqpbx\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.908947 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.908955 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/854575dc-d77d-43fe-a7ff-495ef76ab456-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.912818 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/854575dc-d77d-43fe-a7ff-495ef76ab456-pod-info\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:34 crc kubenswrapper[4946]: I1203 07:10:34.958001 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " pod="openstack/rabbitmq-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.084831 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.084879 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.084906 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.084928 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.084943 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f821218-d2a4-4f97-a924-199c056c7c55-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.084958 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.084982 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.085004 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f821218-d2a4-4f97-a924-199c056c7c55-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.085020 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.085036 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.085060 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5jf6\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-kube-api-access-h5jf6\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186447 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5jf6\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-kube-api-access-h5jf6\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186528 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186555 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186579 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186618 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186646 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f821218-d2a4-4f97-a924-199c056c7c55-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186668 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186709 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186754 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f821218-d2a4-4f97-a924-199c056c7c55-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186779 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.186798 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.187919 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.188207 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.189748 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.190064 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.190479 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.192302 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.194179 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f821218-d2a4-4f97-a924-199c056c7c55-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.201018 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.201369 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.201474 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f821218-d2a4-4f97-a924-199c056c7c55-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.205533 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5jf6\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-kube-api-access-h5jf6\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.229280 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.233548 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.310727 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:10:35 crc kubenswrapper[4946]: I1203 07:10:35.403455 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" event={"ID":"4ba0e22f-5c32-4638-8347-013fa968e479","Type":"ContainerStarted","Data":"a5de8e034bcc3aed831766d620d3865eafa93abda2d9701456283decb0082c29"} Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.100239 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.101403 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.103593 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.104363 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.104923 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nqcgs" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.105662 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.116792 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.116892 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.202537 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.202618 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.202642 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-operator-scripts\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.202684 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdcrr\" (UniqueName: \"kubernetes.io/projected/85593cc4-b6bd-44e9-a908-58f99f0c8756-kube-api-access-zdcrr\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.202714 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-kolla-config\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.202730 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.202769 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-generated\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.202783 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-default\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.303989 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.304219 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.304335 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.304633 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-operator-scripts\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.304843 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdcrr\" (UniqueName: \"kubernetes.io/projected/85593cc4-b6bd-44e9-a908-58f99f0c8756-kube-api-access-zdcrr\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.305338 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-kolla-config\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.305384 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.305456 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-generated\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.305491 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-default\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.307218 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-generated\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.307936 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-kolla-config\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.307974 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-default\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.314997 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-operator-scripts\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.318404 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.322223 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.327450 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdcrr\" (UniqueName: \"kubernetes.io/projected/85593cc4-b6bd-44e9-a908-58f99f0c8756-kube-api-access-zdcrr\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.344335 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " pod="openstack/openstack-galera-0" Dec 03 07:10:36 crc kubenswrapper[4946]: I1203 07:10:36.476475 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.640529 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.650527 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.654997 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.655437 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.655642 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.655821 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-tsrkg" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.694072 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.730767 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.730824 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.730858 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.731085 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.731107 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs29c\" (UniqueName: \"kubernetes.io/projected/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kube-api-access-cs29c\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.731129 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.731143 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.731160 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.832512 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.832819 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs29c\" (UniqueName: \"kubernetes.io/projected/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kube-api-access-cs29c\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.832844 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.832860 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.832880 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.832943 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.832982 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.833010 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.834385 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.835496 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.835850 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.836053 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.837158 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.837358 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.847171 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.850074 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs29c\" (UniqueName: \"kubernetes.io/projected/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kube-api-access-cs29c\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.858169 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.961727 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.962610 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.964913 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-k7pkq" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.964920 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.965453 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.988723 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 07:10:37 crc kubenswrapper[4946]: I1203 07:10:37.995858 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.043004 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-config-data\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.043347 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2k7l\" (UniqueName: \"kubernetes.io/projected/b36490b8-d5db-49c3-97e4-b734feaabd00-kube-api-access-h2k7l\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.043490 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.043727 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-kolla-config\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.043896 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.145579 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-kolla-config\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.145629 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.145697 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-config-data\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.145723 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2k7l\" (UniqueName: \"kubernetes.io/projected/b36490b8-d5db-49c3-97e4-b734feaabd00-kube-api-access-h2k7l\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.145856 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.146310 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-kolla-config\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.146880 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-config-data\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.149338 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.154655 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.162501 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2k7l\" (UniqueName: \"kubernetes.io/projected/b36490b8-d5db-49c3-97e4-b734feaabd00-kube-api-access-h2k7l\") pod \"memcached-0\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " pod="openstack/memcached-0" Dec 03 07:10:38 crc kubenswrapper[4946]: I1203 07:10:38.307611 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 07:10:40 crc kubenswrapper[4946]: I1203 07:10:40.059988 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:10:40 crc kubenswrapper[4946]: I1203 07:10:40.067516 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 07:10:40 crc kubenswrapper[4946]: I1203 07:10:40.070875 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-qrt4q" Dec 03 07:10:40 crc kubenswrapper[4946]: I1203 07:10:40.072213 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:10:40 crc kubenswrapper[4946]: I1203 07:10:40.094314 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxjd7\" (UniqueName: \"kubernetes.io/projected/e0916ede-42cd-4da1-86f0-db41c56c64ec-kube-api-access-cxjd7\") pod \"kube-state-metrics-0\" (UID: \"e0916ede-42cd-4da1-86f0-db41c56c64ec\") " pod="openstack/kube-state-metrics-0" Dec 03 07:10:40 crc kubenswrapper[4946]: I1203 07:10:40.196139 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxjd7\" (UniqueName: \"kubernetes.io/projected/e0916ede-42cd-4da1-86f0-db41c56c64ec-kube-api-access-cxjd7\") pod \"kube-state-metrics-0\" (UID: \"e0916ede-42cd-4da1-86f0-db41c56c64ec\") " pod="openstack/kube-state-metrics-0" Dec 03 07:10:40 crc kubenswrapper[4946]: I1203 07:10:40.225822 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxjd7\" (UniqueName: \"kubernetes.io/projected/e0916ede-42cd-4da1-86f0-db41c56c64ec-kube-api-access-cxjd7\") pod \"kube-state-metrics-0\" (UID: \"e0916ede-42cd-4da1-86f0-db41c56c64ec\") " pod="openstack/kube-state-metrics-0" Dec 03 07:10:40 crc kubenswrapper[4946]: I1203 07:10:40.400093 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.801542 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hc87l"] Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.803208 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hc87l" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.805980 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-2242v" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.806230 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.808438 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.812837 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hc87l"] Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.852102 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-sk56f"] Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.853555 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.858426 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sk56f"] Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.916706 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-log\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.916769 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.916802 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-ovn-controller-tls-certs\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.916949 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-scripts\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917054 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-run\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917103 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-etc-ovs\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917155 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6pnz\" (UniqueName: \"kubernetes.io/projected/f8c1e670-a923-4bf5-a884-23644b2e52aa-kube-api-access-t6pnz\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917187 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8c1e670-a923-4bf5-a884-23644b2e52aa-scripts\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917200 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-log-ovn\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917323 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run-ovn\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917352 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-combined-ca-bundle\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917385 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-lib\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:44 crc kubenswrapper[4946]: I1203 07:10:44.917460 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-244bg\" (UniqueName: \"kubernetes.io/projected/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-kube-api-access-244bg\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019166 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-244bg\" (UniqueName: \"kubernetes.io/projected/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-kube-api-access-244bg\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019257 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-log\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019283 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019302 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-ovn-controller-tls-certs\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019329 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-scripts\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019356 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-run\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019377 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-etc-ovs\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019394 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6pnz\" (UniqueName: \"kubernetes.io/projected/f8c1e670-a923-4bf5-a884-23644b2e52aa-kube-api-access-t6pnz\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019413 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8c1e670-a923-4bf5-a884-23644b2e52aa-scripts\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019431 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-log-ovn\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019451 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-combined-ca-bundle\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019467 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run-ovn\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019486 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-lib\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.019926 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-log\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.020408 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run-ovn\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.020421 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-log-ovn\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.020424 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-etc-ovs\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.021923 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-run\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.022557 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-scripts\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.023864 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8c1e670-a923-4bf5-a884-23644b2e52aa-scripts\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.024641 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.024640 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-lib\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.033402 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-combined-ca-bundle\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.033421 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-ovn-controller-tls-certs\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.038701 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6pnz\" (UniqueName: \"kubernetes.io/projected/f8c1e670-a923-4bf5-a884-23644b2e52aa-kube-api-access-t6pnz\") pod \"ovn-controller-ovs-sk56f\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.038770 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-244bg\" (UniqueName: \"kubernetes.io/projected/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-kube-api-access-244bg\") pod \"ovn-controller-hc87l\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.183206 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hc87l" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.184390 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.681902 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.683419 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.686578 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-kllqn" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.686808 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.686879 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.690039 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.690069 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.697550 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.731658 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.731696 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8260f89c-239d-42e6-b187-c912013d5eca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.731750 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.731774 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvjls\" (UniqueName: \"kubernetes.io/projected/8260f89c-239d-42e6-b187-c912013d5eca-kube-api-access-vvjls\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.731789 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.731923 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.731952 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-config\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.731983 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.834493 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.834541 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8260f89c-239d-42e6-b187-c912013d5eca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.834584 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.834620 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvjls\" (UniqueName: \"kubernetes.io/projected/8260f89c-239d-42e6-b187-c912013d5eca-kube-api-access-vvjls\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.834640 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.834691 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.834722 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-config\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.834803 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.835311 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.835511 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.836899 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8260f89c-239d-42e6-b187-c912013d5eca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.837624 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-config\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.842858 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.844658 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.850513 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.855404 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvjls\" (UniqueName: \"kubernetes.io/projected/8260f89c-239d-42e6-b187-c912013d5eca-kube-api-access-vvjls\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:45 crc kubenswrapper[4946]: I1203 07:10:45.874721 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:46 crc kubenswrapper[4946]: I1203 07:10:46.031134 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.496170 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.497778 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.499555 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.500481 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qxk62" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.501444 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.502077 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.508552 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.565425 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.565472 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.565500 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.565770 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n7dp\" (UniqueName: \"kubernetes.io/projected/7b3d3641-40ca-4391-b7f2-811a85e9c99e-kube-api-access-4n7dp\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.565845 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.565906 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-config\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.566016 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.566058 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.667470 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.667682 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-config\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.667823 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.668753 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-config\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.668950 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.669092 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.669319 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.669435 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.669545 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.669774 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n7dp\" (UniqueName: \"kubernetes.io/projected/7b3d3641-40ca-4391-b7f2-811a85e9c99e-kube-api-access-4n7dp\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.670301 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.671431 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.675554 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.675562 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.675698 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.688247 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n7dp\" (UniqueName: \"kubernetes.io/projected/7b3d3641-40ca-4391-b7f2-811a85e9c99e-kube-api-access-4n7dp\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.700027 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:47 crc kubenswrapper[4946]: I1203 07:10:47.844897 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 07:10:50 crc kubenswrapper[4946]: E1203 07:10:50.689830 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 03 07:10:50 crc kubenswrapper[4946]: E1203 07:10:50.690456 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hdg4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5cd484bb89-9jd5f_openstack(c71804ce-5cc9-40cb-bc49-c1d5620cc034): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:10:50 crc kubenswrapper[4946]: E1203 07:10:50.691649 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" podUID="c71804ce-5cc9-40cb-bc49-c1d5620cc034" Dec 03 07:10:50 crc kubenswrapper[4946]: E1203 07:10:50.723248 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 03 07:10:50 crc kubenswrapper[4946]: E1203 07:10:50.723505 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xv2pm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-567c455747-jh9sl_openstack(f22bc63b-9a11-4393-8976-15a2eb992ee2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:10:50 crc kubenswrapper[4946]: E1203 07:10:50.725159 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-567c455747-jh9sl" podUID="f22bc63b-9a11-4393-8976-15a2eb992ee2" Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.227472 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.391138 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.575224 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345","Type":"ContainerStarted","Data":"42caf5eaffec889c4853100be6c5a46874c3730f08900043c8dfa7757d66c6df"} Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.577214 4946 generic.go:334] "Generic (PLEG): container finished" podID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" containerID="f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94" exitCode=0 Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.577285 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" event={"ID":"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb","Type":"ContainerDied","Data":"f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94"} Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.580379 4946 generic.go:334] "Generic (PLEG): container finished" podID="4ba0e22f-5c32-4638-8347-013fa968e479" containerID="13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69" exitCode=0 Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.580475 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" event={"ID":"4ba0e22f-5c32-4638-8347-013fa968e479","Type":"ContainerDied","Data":"13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69"} Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.583205 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2f821218-d2a4-4f97-a924-199c056c7c55","Type":"ContainerStarted","Data":"85a4854921c2c14df675d2f527a4f30ec3f96f81ee2ccb5e5c48d858434366c3"} Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.751186 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.780349 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.800337 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.816330 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hc87l"] Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.823590 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.905792 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 07:10:51 crc kubenswrapper[4946]: I1203 07:10:51.994962 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.063967 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71804ce-5cc9-40cb-bc49-c1d5620cc034-config\") pod \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\" (UID: \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\") " Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.064073 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdg4c\" (UniqueName: \"kubernetes.io/projected/c71804ce-5cc9-40cb-bc49-c1d5620cc034-kube-api-access-hdg4c\") pod \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\" (UID: \"c71804ce-5cc9-40cb-bc49-c1d5620cc034\") " Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.064607 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c71804ce-5cc9-40cb-bc49-c1d5620cc034-config" (OuterVolumeSpecName: "config") pod "c71804ce-5cc9-40cb-bc49-c1d5620cc034" (UID: "c71804ce-5cc9-40cb-bc49-c1d5620cc034"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.069397 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c71804ce-5cc9-40cb-bc49-c1d5620cc034-kube-api-access-hdg4c" (OuterVolumeSpecName: "kube-api-access-hdg4c") pod "c71804ce-5cc9-40cb-bc49-c1d5620cc034" (UID: "c71804ce-5cc9-40cb-bc49-c1d5620cc034"). InnerVolumeSpecName "kube-api-access-hdg4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.096505 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.166178 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-dns-svc\") pod \"f22bc63b-9a11-4393-8976-15a2eb992ee2\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.166364 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-config\") pod \"f22bc63b-9a11-4393-8976-15a2eb992ee2\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.166460 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv2pm\" (UniqueName: \"kubernetes.io/projected/f22bc63b-9a11-4393-8976-15a2eb992ee2-kube-api-access-xv2pm\") pod \"f22bc63b-9a11-4393-8976-15a2eb992ee2\" (UID: \"f22bc63b-9a11-4393-8976-15a2eb992ee2\") " Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.166517 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f22bc63b-9a11-4393-8976-15a2eb992ee2" (UID: "f22bc63b-9a11-4393-8976-15a2eb992ee2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.166785 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-config" (OuterVolumeSpecName: "config") pod "f22bc63b-9a11-4393-8976-15a2eb992ee2" (UID: "f22bc63b-9a11-4393-8976-15a2eb992ee2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.167054 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.167072 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71804ce-5cc9-40cb-bc49-c1d5620cc034-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.167082 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdg4c\" (UniqueName: \"kubernetes.io/projected/c71804ce-5cc9-40cb-bc49-c1d5620cc034-kube-api-access-hdg4c\") on node \"crc\" DevicePath \"\"" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.167092 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f22bc63b-9a11-4393-8976-15a2eb992ee2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.170752 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f22bc63b-9a11-4393-8976-15a2eb992ee2-kube-api-access-xv2pm" (OuterVolumeSpecName: "kube-api-access-xv2pm") pod "f22bc63b-9a11-4393-8976-15a2eb992ee2" (UID: "f22bc63b-9a11-4393-8976-15a2eb992ee2"). InnerVolumeSpecName "kube-api-access-xv2pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.269397 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv2pm\" (UniqueName: \"kubernetes.io/projected/f22bc63b-9a11-4393-8976-15a2eb992ee2-kube-api-access-xv2pm\") on node \"crc\" DevicePath \"\"" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.598106 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.598149 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-9jd5f" event={"ID":"c71804ce-5cc9-40cb-bc49-c1d5620cc034","Type":"ContainerDied","Data":"77b296145fdf825fd5e7c7b3203b0a6efad7f85b71b0b4ce3ac565b07a68980f"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.600246 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b3d3641-40ca-4391-b7f2-811a85e9c99e","Type":"ContainerStarted","Data":"707e8b7baeee32b7b585df2c1a766119a5228d390ce4b3b00eab5cc7ffd10a48"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.603379 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" event={"ID":"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb","Type":"ContainerStarted","Data":"4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.603476 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.604682 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hc87l" event={"ID":"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d","Type":"ContainerStarted","Data":"484cecebdb5b1d74f73d5421d9f200478f97e3f71bdc33a2c78f5251c475aada"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.605816 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-jh9sl" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.605800 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-jh9sl" event={"ID":"f22bc63b-9a11-4393-8976-15a2eb992ee2","Type":"ContainerDied","Data":"bbb75635389a2b5728a1b3fa784ebbc1d3df794e85ca6893c358cd79bf279f6a"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.607024 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"85593cc4-b6bd-44e9-a908-58f99f0c8756","Type":"ContainerStarted","Data":"465c4b27ada46636f29c78d4626fd8ed46110914b69a336eb96d1ee7a96f4b39"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.607980 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e0916ede-42cd-4da1-86f0-db41c56c64ec","Type":"ContainerStarted","Data":"cbd535d9e8ea1233707a0e8fcb2b67b8c22be494735759541aa75cfa4974e045"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.610474 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" event={"ID":"4ba0e22f-5c32-4638-8347-013fa968e479","Type":"ContainerStarted","Data":"162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.610611 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.611807 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"854575dc-d77d-43fe-a7ff-495ef76ab456","Type":"ContainerStarted","Data":"14aafb403e6899be5493858d2c6fb52724774ca8c09d0f1db1e4099cffb954ff"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.613622 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b36490b8-d5db-49c3-97e4-b734feaabd00","Type":"ContainerStarted","Data":"e9075da2f08d4d16898fc1dabf6593b7ad42915749b7c7fe8fea116b6a042271"} Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.628423 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" podStartSLOduration=2.998055746 podStartE2EDuration="19.628405502s" podCreationTimestamp="2025-12-03 07:10:33 +0000 UTC" firstStartedPulling="2025-12-03 07:10:34.262603648 +0000 UTC m=+1227.059293757" lastFinishedPulling="2025-12-03 07:10:50.892953404 +0000 UTC m=+1243.689643513" observedRunningTime="2025-12-03 07:10:52.618837857 +0000 UTC m=+1245.415527966" watchObservedRunningTime="2025-12-03 07:10:52.628405502 +0000 UTC m=+1245.425095611" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.653476 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" podStartSLOduration=3.335036471 podStartE2EDuration="19.653453281s" podCreationTimestamp="2025-12-03 07:10:33 +0000 UTC" firstStartedPulling="2025-12-03 07:10:34.559876043 +0000 UTC m=+1227.356566152" lastFinishedPulling="2025-12-03 07:10:50.878292853 +0000 UTC m=+1243.674982962" observedRunningTime="2025-12-03 07:10:52.642475758 +0000 UTC m=+1245.439165867" watchObservedRunningTime="2025-12-03 07:10:52.653453281 +0000 UTC m=+1245.450143380" Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.682136 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-jh9sl"] Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.695006 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-567c455747-jh9sl"] Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.710785 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-9jd5f"] Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.718034 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-9jd5f"] Dec 03 07:10:52 crc kubenswrapper[4946]: I1203 07:10:52.886430 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 07:10:53 crc kubenswrapper[4946]: I1203 07:10:53.018596 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sk56f"] Dec 03 07:10:53 crc kubenswrapper[4946]: I1203 07:10:53.607161 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c71804ce-5cc9-40cb-bc49-c1d5620cc034" path="/var/lib/kubelet/pods/c71804ce-5cc9-40cb-bc49-c1d5620cc034/volumes" Dec 03 07:10:53 crc kubenswrapper[4946]: I1203 07:10:53.607614 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f22bc63b-9a11-4393-8976-15a2eb992ee2" path="/var/lib/kubelet/pods/f22bc63b-9a11-4393-8976-15a2eb992ee2/volumes" Dec 03 07:10:53 crc kubenswrapper[4946]: I1203 07:10:53.625595 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8260f89c-239d-42e6-b187-c912013d5eca","Type":"ContainerStarted","Data":"cd6cd0524b7074097a1a5912f376575963143b3d36d4647375c6922450a76242"} Dec 03 07:10:54 crc kubenswrapper[4946]: I1203 07:10:54.646293 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sk56f" event={"ID":"f8c1e670-a923-4bf5-a884-23644b2e52aa","Type":"ContainerStarted","Data":"8191b4220a9e3c19d14db76a25889dc5cc3fdcb424688a3510ab0ba368e40f2c"} Dec 03 07:10:58 crc kubenswrapper[4946]: I1203 07:10:58.780036 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:10:59 crc kubenswrapper[4946]: I1203 07:10:59.085387 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:10:59 crc kubenswrapper[4946]: I1203 07:10:59.133880 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-xnm28"] Dec 03 07:10:59 crc kubenswrapper[4946]: I1203 07:10:59.694847 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" podUID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" containerName="dnsmasq-dns" containerID="cri-o://4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0" gracePeriod=10 Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.470189 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.532557 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz7j2\" (UniqueName: \"kubernetes.io/projected/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-kube-api-access-fz7j2\") pod \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.532666 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-dns-svc\") pod \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.532726 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-config\") pod \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\" (UID: \"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb\") " Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.548698 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-kube-api-access-fz7j2" (OuterVolumeSpecName: "kube-api-access-fz7j2") pod "8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" (UID: "8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb"). InnerVolumeSpecName "kube-api-access-fz7j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.603421 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-config" (OuterVolumeSpecName: "config") pod "8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" (UID: "8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.612722 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" (UID: "8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.636764 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz7j2\" (UniqueName: \"kubernetes.io/projected/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-kube-api-access-fz7j2\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.636795 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.636806 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.703560 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b3d3641-40ca-4391-b7f2-811a85e9c99e","Type":"ContainerStarted","Data":"306f0aa54817d7e3ac475556197a16b148ab3536c5c0e1a6f77075f7234d7b98"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.705515 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345","Type":"ContainerStarted","Data":"ffed1397d4af3d4cd73aff15e45d307e90a8f9c3d60ab75e2f08709cf0802df3"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.707908 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hc87l" event={"ID":"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d","Type":"ContainerStarted","Data":"8293f410bee38eef4e13c919564b09535995b3ce313122377c14666333bd220b"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.708045 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-hc87l" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.709903 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8260f89c-239d-42e6-b187-c912013d5eca","Type":"ContainerStarted","Data":"5b78ba643d512f9707208df0f4f8fcbbe76854d669c9d51da5ef8396964bb755"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.711816 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b36490b8-d5db-49c3-97e4-b734feaabd00","Type":"ContainerStarted","Data":"40db7579ead1028473e650e6ae3b6bc768d742221d67f47bd5ce52721da60990"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.711915 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.713424 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"85593cc4-b6bd-44e9-a908-58f99f0c8756","Type":"ContainerStarted","Data":"2415dc5d63b46387779f532cd38fe4a49aee041aa819437a99561449e7e5821c"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.715326 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sk56f" event={"ID":"f8c1e670-a923-4bf5-a884-23644b2e52aa","Type":"ContainerStarted","Data":"b26d19402ccd1d5d3bfc46b1bae7e60dec0a258f0bebe9c9185a23fec83336f4"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.716924 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e0916ede-42cd-4da1-86f0-db41c56c64ec","Type":"ContainerStarted","Data":"41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.716989 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.718540 4946 generic.go:334] "Generic (PLEG): container finished" podID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" containerID="4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0" exitCode=0 Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.718574 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" event={"ID":"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb","Type":"ContainerDied","Data":"4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.718590 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" event={"ID":"8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb","Type":"ContainerDied","Data":"c97d21d0acf1eb760e9738395def00851124acb199f8c89252d63bc5856f609a"} Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.718606 4946 scope.go:117] "RemoveContainer" containerID="4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.718688 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-xnm28" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.759439 4946 scope.go:117] "RemoveContainer" containerID="f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.766305 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=16.517935644 podStartE2EDuration="23.766279362s" podCreationTimestamp="2025-12-03 07:10:37 +0000 UTC" firstStartedPulling="2025-12-03 07:10:51.775718045 +0000 UTC m=+1244.572408154" lastFinishedPulling="2025-12-03 07:10:59.024061763 +0000 UTC m=+1251.820751872" observedRunningTime="2025-12-03 07:11:00.748550298 +0000 UTC m=+1253.545240407" watchObservedRunningTime="2025-12-03 07:11:00.766279362 +0000 UTC m=+1253.562969471" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.794961 4946 scope.go:117] "RemoveContainer" containerID="4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0" Dec 03 07:11:00 crc kubenswrapper[4946]: E1203 07:11:00.795442 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0\": container with ID starting with 4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0 not found: ID does not exist" containerID="4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.795476 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0"} err="failed to get container status \"4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0\": rpc error: code = NotFound desc = could not find container \"4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0\": container with ID starting with 4c8d643e8f7b1385657ea7f47c9c6d879f055af86249c4dffa11f3d0a56afaa0 not found: ID does not exist" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.795508 4946 scope.go:117] "RemoveContainer" containerID="f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94" Dec 03 07:11:00 crc kubenswrapper[4946]: E1203 07:11:00.795890 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94\": container with ID starting with f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94 not found: ID does not exist" containerID="f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.795932 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94"} err="failed to get container status \"f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94\": rpc error: code = NotFound desc = could not find container \"f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94\": container with ID starting with f19904d68cd567e12ba7e973baad8c5774091f22dec6acdae8df27a873158d94 not found: ID does not exist" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.829898 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-hc87l" podStartSLOduration=9.00639558 podStartE2EDuration="16.829882261s" podCreationTimestamp="2025-12-03 07:10:44 +0000 UTC" firstStartedPulling="2025-12-03 07:10:51.786033401 +0000 UTC m=+1244.582723510" lastFinishedPulling="2025-12-03 07:10:59.609520062 +0000 UTC m=+1252.406210191" observedRunningTime="2025-12-03 07:11:00.829654635 +0000 UTC m=+1253.626344744" watchObservedRunningTime="2025-12-03 07:11:00.829882261 +0000 UTC m=+1253.626572370" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.855884 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=12.359760936 podStartE2EDuration="20.855865515s" podCreationTimestamp="2025-12-03 07:10:40 +0000 UTC" firstStartedPulling="2025-12-03 07:10:51.792241566 +0000 UTC m=+1244.588931675" lastFinishedPulling="2025-12-03 07:11:00.288346155 +0000 UTC m=+1253.085036254" observedRunningTime="2025-12-03 07:11:00.85193362 +0000 UTC m=+1253.648623739" watchObservedRunningTime="2025-12-03 07:11:00.855865515 +0000 UTC m=+1253.652555614" Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.864614 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-xnm28"] Dec 03 07:11:00 crc kubenswrapper[4946]: I1203 07:11:00.869853 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-xnm28"] Dec 03 07:11:01 crc kubenswrapper[4946]: I1203 07:11:01.604948 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" path="/var/lib/kubelet/pods/8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb/volumes" Dec 03 07:11:01 crc kubenswrapper[4946]: I1203 07:11:01.728831 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"854575dc-d77d-43fe-a7ff-495ef76ab456","Type":"ContainerStarted","Data":"5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78"} Dec 03 07:11:01 crc kubenswrapper[4946]: I1203 07:11:01.730612 4946 generic.go:334] "Generic (PLEG): container finished" podID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerID="b26d19402ccd1d5d3bfc46b1bae7e60dec0a258f0bebe9c9185a23fec83336f4" exitCode=0 Dec 03 07:11:01 crc kubenswrapper[4946]: I1203 07:11:01.730910 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sk56f" event={"ID":"f8c1e670-a923-4bf5-a884-23644b2e52aa","Type":"ContainerDied","Data":"b26d19402ccd1d5d3bfc46b1bae7e60dec0a258f0bebe9c9185a23fec83336f4"} Dec 03 07:11:01 crc kubenswrapper[4946]: I1203 07:11:01.735066 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2f821218-d2a4-4f97-a924-199c056c7c55","Type":"ContainerStarted","Data":"15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d"} Dec 03 07:11:02 crc kubenswrapper[4946]: I1203 07:11:02.743580 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sk56f" event={"ID":"f8c1e670-a923-4bf5-a884-23644b2e52aa","Type":"ContainerStarted","Data":"2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732"} Dec 03 07:11:03 crc kubenswrapper[4946]: I1203 07:11:03.772981 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sk56f" event={"ID":"f8c1e670-a923-4bf5-a884-23644b2e52aa","Type":"ContainerStarted","Data":"019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b"} Dec 03 07:11:03 crc kubenswrapper[4946]: I1203 07:11:03.774445 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:11:03 crc kubenswrapper[4946]: I1203 07:11:03.777221 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b3d3641-40ca-4391-b7f2-811a85e9c99e","Type":"ContainerStarted","Data":"dbf1d62862ef8e193ae4fb4237e071cfc9defb2902e956f02581df96ce0e9858"} Dec 03 07:11:03 crc kubenswrapper[4946]: I1203 07:11:03.798485 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8260f89c-239d-42e6-b187-c912013d5eca","Type":"ContainerStarted","Data":"bbdd199402f5e0e7ad481bf616fe43d3edcdb9ecdbe811804101a2daa767e736"} Dec 03 07:11:03 crc kubenswrapper[4946]: I1203 07:11:03.814331 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-sk56f" podStartSLOduration=13.945129494 podStartE2EDuration="19.814304391s" podCreationTimestamp="2025-12-03 07:10:44 +0000 UTC" firstStartedPulling="2025-12-03 07:10:54.196807198 +0000 UTC m=+1246.993497307" lastFinishedPulling="2025-12-03 07:11:00.065982095 +0000 UTC m=+1252.862672204" observedRunningTime="2025-12-03 07:11:03.802974808 +0000 UTC m=+1256.599664987" watchObservedRunningTime="2025-12-03 07:11:03.814304391 +0000 UTC m=+1256.610994540" Dec 03 07:11:03 crc kubenswrapper[4946]: I1203 07:11:03.850284 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=6.569756505 podStartE2EDuration="17.850261881s" podCreationTimestamp="2025-12-03 07:10:46 +0000 UTC" firstStartedPulling="2025-12-03 07:10:51.940452035 +0000 UTC m=+1244.737142144" lastFinishedPulling="2025-12-03 07:11:03.220957401 +0000 UTC m=+1256.017647520" observedRunningTime="2025-12-03 07:11:03.841425345 +0000 UTC m=+1256.638115484" watchObservedRunningTime="2025-12-03 07:11:03.850261881 +0000 UTC m=+1256.646952000" Dec 03 07:11:03 crc kubenswrapper[4946]: I1203 07:11:03.872603 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.540621179 podStartE2EDuration="19.872573217s" podCreationTimestamp="2025-12-03 07:10:44 +0000 UTC" firstStartedPulling="2025-12-03 07:10:52.910433996 +0000 UTC m=+1245.707124105" lastFinishedPulling="2025-12-03 07:11:03.242386014 +0000 UTC m=+1256.039076143" observedRunningTime="2025-12-03 07:11:03.866288269 +0000 UTC m=+1256.662978388" watchObservedRunningTime="2025-12-03 07:11:03.872573217 +0000 UTC m=+1256.669263346" Dec 03 07:11:04 crc kubenswrapper[4946]: I1203 07:11:04.032837 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 07:11:04 crc kubenswrapper[4946]: I1203 07:11:04.075184 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 07:11:04 crc kubenswrapper[4946]: I1203 07:11:04.813653 4946 generic.go:334] "Generic (PLEG): container finished" podID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" containerID="ffed1397d4af3d4cd73aff15e45d307e90a8f9c3d60ab75e2f08709cf0802df3" exitCode=0 Dec 03 07:11:04 crc kubenswrapper[4946]: I1203 07:11:04.814060 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345","Type":"ContainerDied","Data":"ffed1397d4af3d4cd73aff15e45d307e90a8f9c3d60ab75e2f08709cf0802df3"} Dec 03 07:11:04 crc kubenswrapper[4946]: I1203 07:11:04.818551 4946 generic.go:334] "Generic (PLEG): container finished" podID="85593cc4-b6bd-44e9-a908-58f99f0c8756" containerID="2415dc5d63b46387779f532cd38fe4a49aee041aa819437a99561449e7e5821c" exitCode=0 Dec 03 07:11:04 crc kubenswrapper[4946]: I1203 07:11:04.818926 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"85593cc4-b6bd-44e9-a908-58f99f0c8756","Type":"ContainerDied","Data":"2415dc5d63b46387779f532cd38fe4a49aee041aa819437a99561449e7e5821c"} Dec 03 07:11:04 crc kubenswrapper[4946]: I1203 07:11:04.820543 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 07:11:04 crc kubenswrapper[4946]: I1203 07:11:04.820592 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:11:05 crc kubenswrapper[4946]: I1203 07:11:05.828978 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"85593cc4-b6bd-44e9-a908-58f99f0c8756","Type":"ContainerStarted","Data":"a625c10f25213752b4f3ffbe138270f09d8ca2e6105053ba6f35c9e4d430ce6e"} Dec 03 07:11:05 crc kubenswrapper[4946]: I1203 07:11:05.834085 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345","Type":"ContainerStarted","Data":"f019b96d17fb751f8b25510337f077e065922ffd4d5f69477d62724118532891"} Dec 03 07:11:05 crc kubenswrapper[4946]: I1203 07:11:05.846027 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 07:11:05 crc kubenswrapper[4946]: I1203 07:11:05.856096 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=23.246609605 podStartE2EDuration="30.85606767s" podCreationTimestamp="2025-12-03 07:10:35 +0000 UTC" firstStartedPulling="2025-12-03 07:10:51.78040805 +0000 UTC m=+1244.577098159" lastFinishedPulling="2025-12-03 07:10:59.389866115 +0000 UTC m=+1252.186556224" observedRunningTime="2025-12-03 07:11:05.854814687 +0000 UTC m=+1258.651504806" watchObservedRunningTime="2025-12-03 07:11:05.85606767 +0000 UTC m=+1258.652757789" Dec 03 07:11:05 crc kubenswrapper[4946]: I1203 07:11:05.892025 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.73764507 podStartE2EDuration="29.89200397s" podCreationTimestamp="2025-12-03 07:10:36 +0000 UTC" firstStartedPulling="2025-12-03 07:10:51.398012836 +0000 UTC m=+1244.194702955" lastFinishedPulling="2025-12-03 07:10:59.552371746 +0000 UTC m=+1252.349061855" observedRunningTime="2025-12-03 07:11:05.873340932 +0000 UTC m=+1258.670031111" watchObservedRunningTime="2025-12-03 07:11:05.89200397 +0000 UTC m=+1258.688694079" Dec 03 07:11:05 crc kubenswrapper[4946]: I1203 07:11:05.892282 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 07:11:05 crc kubenswrapper[4946]: I1203 07:11:05.925355 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.171231 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c67bcdbf5-sdq9h"] Dec 03 07:11:06 crc kubenswrapper[4946]: E1203 07:11:06.171554 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" containerName="dnsmasq-dns" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.171570 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" containerName="dnsmasq-dns" Dec 03 07:11:06 crc kubenswrapper[4946]: E1203 07:11:06.171602 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" containerName="init" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.171608 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" containerName="init" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.171762 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8072e36c-6ff2-49fe-bcce-99e7ebd2f1bb" containerName="dnsmasq-dns" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.172589 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.175444 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.192089 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c67bcdbf5-sdq9h"] Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.260863 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-fh88p"] Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.262164 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.265180 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.283656 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fh88p"] Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.353181 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-ovsdbserver-nb\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.353229 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-dns-svc\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.353247 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-config\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.353318 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9qs6\" (UniqueName: \"kubernetes.io/projected/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-kube-api-access-l9qs6\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.404471 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c67bcdbf5-sdq9h"] Dec 03 07:11:06 crc kubenswrapper[4946]: E1203 07:11:06.405350 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-l9qs6 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" podUID="fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.432304 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-984c76dd7-rp5sp"] Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.433504 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.435054 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.442232 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-984c76dd7-rp5sp"] Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454301 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-dns-svc\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454351 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-config\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454402 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9qs6\" (UniqueName: \"kubernetes.io/projected/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-kube-api-access-l9qs6\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454446 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovs-rundir\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454480 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-combined-ca-bundle\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454520 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454552 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpdsw\" (UniqueName: \"kubernetes.io/projected/df3b0afc-1419-44ce-89f7-b1aca189e145-kube-api-access-jpdsw\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454580 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovn-rundir\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454629 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df3b0afc-1419-44ce-89f7-b1aca189e145-config\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.454670 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-ovsdbserver-nb\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.455266 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-dns-svc\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.455685 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-ovsdbserver-nb\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.456371 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-config\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.476754 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9qs6\" (UniqueName: \"kubernetes.io/projected/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-kube-api-access-l9qs6\") pod \"dnsmasq-dns-6c67bcdbf5-sdq9h\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.477168 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.477200 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556328 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-sb\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556385 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovs-rundir\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556417 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-combined-ca-bundle\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556435 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-config\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556649 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-nb\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556795 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556840 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovs-rundir\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556856 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpdsw\" (UniqueName: \"kubernetes.io/projected/df3b0afc-1419-44ce-89f7-b1aca189e145-kube-api-access-jpdsw\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.556914 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovn-rundir\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.557011 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgpnz\" (UniqueName: \"kubernetes.io/projected/c33d80bb-9824-45ed-8945-11e459e54f25-kube-api-access-fgpnz\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.557074 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df3b0afc-1419-44ce-89f7-b1aca189e145-config\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.557160 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovn-rundir\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.557276 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-dns-svc\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.557792 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df3b0afc-1419-44ce-89f7-b1aca189e145-config\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.560469 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-combined-ca-bundle\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.560971 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.585071 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpdsw\" (UniqueName: \"kubernetes.io/projected/df3b0afc-1419-44ce-89f7-b1aca189e145-kube-api-access-jpdsw\") pod \"ovn-controller-metrics-fh88p\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.658893 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgpnz\" (UniqueName: \"kubernetes.io/projected/c33d80bb-9824-45ed-8945-11e459e54f25-kube-api-access-fgpnz\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.658982 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-dns-svc\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.659032 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-sb\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.659061 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-config\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.659092 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-nb\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.660073 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-config\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.660076 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-nb\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.660229 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-sb\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.660633 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-dns-svc\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.680665 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgpnz\" (UniqueName: \"kubernetes.io/projected/c33d80bb-9824-45ed-8945-11e459e54f25-kube-api-access-fgpnz\") pod \"dnsmasq-dns-984c76dd7-rp5sp\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.759863 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.843640 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.844735 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.859330 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.882201 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.890527 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.962954 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-ovsdbserver-nb\") pod \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.963176 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-config\") pod \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.963233 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-dns-svc\") pod \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.963278 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9qs6\" (UniqueName: \"kubernetes.io/projected/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-kube-api-access-l9qs6\") pod \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\" (UID: \"fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1\") " Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.963616 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1" (UID: "fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.963992 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.964911 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-config" (OuterVolumeSpecName: "config") pod "fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1" (UID: "fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.965882 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1" (UID: "fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:06 crc kubenswrapper[4946]: I1203 07:11:06.967603 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-kube-api-access-l9qs6" (OuterVolumeSpecName: "kube-api-access-l9qs6") pod "fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1" (UID: "fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1"). InnerVolumeSpecName "kube-api-access-l9qs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.072267 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.072329 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.072340 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9qs6\" (UniqueName: \"kubernetes.io/projected/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1-kube-api-access-l9qs6\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.078798 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.084108 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.088090 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.088260 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-zw6tj" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.088519 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.088554 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.128781 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.175503 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.175836 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrpfz\" (UniqueName: \"kubernetes.io/projected/80557a83-debb-4e3a-a2d0-bb5ac72d824c-kube-api-access-mrpfz\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.175855 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.175890 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-scripts\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.175932 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.175954 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-config\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.175972 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.254876 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-984c76dd7-rp5sp"] Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.287963 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.288022 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.288187 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrpfz\" (UniqueName: \"kubernetes.io/projected/80557a83-debb-4e3a-a2d0-bb5ac72d824c-kube-api-access-mrpfz\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.288225 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.288271 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-scripts\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.288320 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.288351 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-config\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.289445 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.289697 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-config\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.290237 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-scripts\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.294553 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.308067 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fh88p"] Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.309384 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.309604 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrpfz\" (UniqueName: \"kubernetes.io/projected/80557a83-debb-4e3a-a2d0-bb5ac72d824c-kube-api-access-mrpfz\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.309873 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.407649 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.851961 4946 generic.go:334] "Generic (PLEG): container finished" podID="c33d80bb-9824-45ed-8945-11e459e54f25" containerID="e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685" exitCode=0 Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.852019 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" event={"ID":"c33d80bb-9824-45ed-8945-11e459e54f25","Type":"ContainerDied","Data":"e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685"} Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.852478 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" event={"ID":"c33d80bb-9824-45ed-8945-11e459e54f25","Type":"ContainerStarted","Data":"2f1fb8bfc48d0520ba0c7797fe63271efe3cd7e05503f6e90b0634031e1b0d4d"} Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.863150 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c67bcdbf5-sdq9h" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.863891 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fh88p" event={"ID":"df3b0afc-1419-44ce-89f7-b1aca189e145","Type":"ContainerStarted","Data":"2eb06361aba92c8a4af213131c660c362c2c6949de5340c6fe246ad8413ed5fd"} Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.863919 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fh88p" event={"ID":"df3b0afc-1419-44ce-89f7-b1aca189e145","Type":"ContainerStarted","Data":"2dbfac30a23b90732cb1205a995f0464563b345021e6f412353da20ae2cb4121"} Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.863933 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 07:11:07 crc kubenswrapper[4946]: W1203 07:11:07.870722 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80557a83_debb_4e3a_a2d0_bb5ac72d824c.slice/crio-5ef0b652acc68849990842cb7fa135cf01021eef075d817329b9cb84a532cf80 WatchSource:0}: Error finding container 5ef0b652acc68849990842cb7fa135cf01021eef075d817329b9cb84a532cf80: Status 404 returned error can't find the container with id 5ef0b652acc68849990842cb7fa135cf01021eef075d817329b9cb84a532cf80 Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.906363 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-fh88p" podStartSLOduration=1.9063427480000001 podStartE2EDuration="1.906342748s" podCreationTimestamp="2025-12-03 07:11:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:07.905396813 +0000 UTC m=+1260.702086922" watchObservedRunningTime="2025-12-03 07:11:07.906342748 +0000 UTC m=+1260.703032857" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.989999 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 07:11:07 crc kubenswrapper[4946]: I1203 07:11:07.990058 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 07:11:08 crc kubenswrapper[4946]: I1203 07:11:08.071594 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c67bcdbf5-sdq9h"] Dec 03 07:11:08 crc kubenswrapper[4946]: I1203 07:11:08.071650 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c67bcdbf5-sdq9h"] Dec 03 07:11:08 crc kubenswrapper[4946]: I1203 07:11:08.309995 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 07:11:08 crc kubenswrapper[4946]: I1203 07:11:08.872037 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"80557a83-debb-4e3a-a2d0-bb5ac72d824c","Type":"ContainerStarted","Data":"5ef0b652acc68849990842cb7fa135cf01021eef075d817329b9cb84a532cf80"} Dec 03 07:11:09 crc kubenswrapper[4946]: I1203 07:11:09.602988 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1" path="/var/lib/kubelet/pods/fbd38bd7-f065-4b2e-88e3-de5a8f4b96f1/volumes" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.405973 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.509843 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-984c76dd7-rp5sp"] Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.541946 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784d65c867-79wht"] Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.544167 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.555572 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784d65c867-79wht"] Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.643569 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-sb\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.643630 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-nb\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.643672 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-config\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.643707 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4d27\" (UniqueName: \"kubernetes.io/projected/48843445-ca95-4a91-bf4b-05c16ebc945c-kube-api-access-f4d27\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.643758 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-dns-svc\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.745424 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-sb\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.745476 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-nb\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.745519 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-config\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.745556 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4d27\" (UniqueName: \"kubernetes.io/projected/48843445-ca95-4a91-bf4b-05c16ebc945c-kube-api-access-f4d27\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.745587 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-dns-svc\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.746402 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-sb\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.746411 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-dns-svc\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.746617 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-config\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.747065 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-nb\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.763691 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4d27\" (UniqueName: \"kubernetes.io/projected/48843445-ca95-4a91-bf4b-05c16ebc945c-kube-api-access-f4d27\") pod \"dnsmasq-dns-784d65c867-79wht\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:10 crc kubenswrapper[4946]: I1203 07:11:10.868560 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.326848 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784d65c867-79wht"] Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.629372 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.639486 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.641898 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.642147 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.642199 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-zkd4w" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.642309 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.663353 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.764264 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.764329 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.764380 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-lock\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.764414 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-cache\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.764434 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l7qz\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-kube-api-access-5l7qz\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.865381 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-lock\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.865435 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-cache\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.865459 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l7qz\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-kube-api-access-5l7qz\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.865513 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.865550 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.865889 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: E1203 07:11:11.866027 4946 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 07:11:11 crc kubenswrapper[4946]: E1203 07:11:11.866085 4946 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 07:11:11 crc kubenswrapper[4946]: E1203 07:11:11.866176 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift podName:54cbc35f-afb5-4901-a64f-5e08fee6fd22 nodeName:}" failed. No retries permitted until 2025-12-03 07:11:12.366148023 +0000 UTC m=+1265.162838172 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift") pod "swift-storage-0" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22") : configmap "swift-ring-files" not found Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.866685 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-cache\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.866873 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-lock\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.900534 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l7qz\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-kube-api-access-5l7qz\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.908389 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:11 crc kubenswrapper[4946]: I1203 07:11:11.910677 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784d65c867-79wht" event={"ID":"48843445-ca95-4a91-bf4b-05c16ebc945c","Type":"ContainerStarted","Data":"6fb8b1393fdd7fed0333560564387f168d4b885a114eb568fb4c475a4ea1bec6"} Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.221557 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-s624s"] Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.223360 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.227940 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.228910 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.231674 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-s624s"] Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.232100 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.272389 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-scripts\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.272606 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-combined-ca-bundle\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.272685 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-ring-data-devices\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.272816 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkpbx\" (UniqueName: \"kubernetes.io/projected/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-kube-api-access-tkpbx\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.272851 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-etc-swift\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.272888 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-dispersionconf\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.272988 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-swiftconf\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.374588 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-ring-data-devices\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.374673 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkpbx\" (UniqueName: \"kubernetes.io/projected/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-kube-api-access-tkpbx\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.374704 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-etc-swift\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.374755 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-dispersionconf\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.374794 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-swiftconf\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.374833 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-scripts\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.374857 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-combined-ca-bundle\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.374886 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:12 crc kubenswrapper[4946]: E1203 07:11:12.375041 4946 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 07:11:12 crc kubenswrapper[4946]: E1203 07:11:12.375057 4946 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 07:11:12 crc kubenswrapper[4946]: E1203 07:11:12.375103 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift podName:54cbc35f-afb5-4901-a64f-5e08fee6fd22 nodeName:}" failed. No retries permitted until 2025-12-03 07:11:13.375086887 +0000 UTC m=+1266.171776996 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift") pod "swift-storage-0" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22") : configmap "swift-ring-files" not found Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.375873 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-etc-swift\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.375961 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-ring-data-devices\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.376850 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-scripts\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.379978 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-dispersionconf\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.380865 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-combined-ca-bundle\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.392679 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-swiftconf\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.393241 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkpbx\" (UniqueName: \"kubernetes.io/projected/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-kube-api-access-tkpbx\") pod \"swift-ring-rebalance-s624s\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.556144 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:12 crc kubenswrapper[4946]: I1203 07:11:12.923631 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" event={"ID":"c33d80bb-9824-45ed-8945-11e459e54f25","Type":"ContainerStarted","Data":"621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1"} Dec 03 07:11:13 crc kubenswrapper[4946]: I1203 07:11:13.123214 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-s624s"] Dec 03 07:11:13 crc kubenswrapper[4946]: I1203 07:11:13.450587 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:13 crc kubenswrapper[4946]: E1203 07:11:13.450834 4946 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 07:11:13 crc kubenswrapper[4946]: E1203 07:11:13.450854 4946 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 07:11:13 crc kubenswrapper[4946]: E1203 07:11:13.450904 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift podName:54cbc35f-afb5-4901-a64f-5e08fee6fd22 nodeName:}" failed. No retries permitted until 2025-12-03 07:11:15.450887504 +0000 UTC m=+1268.247577623 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift") pod "swift-storage-0" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22") : configmap "swift-ring-files" not found Dec 03 07:11:13 crc kubenswrapper[4946]: I1203 07:11:13.933015 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-s624s" event={"ID":"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a","Type":"ContainerStarted","Data":"d95fa87979592cadaff668dca31906babc196cee1f928e7e0dd311313888e9b1"} Dec 03 07:11:13 crc kubenswrapper[4946]: I1203 07:11:13.933183 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" podUID="c33d80bb-9824-45ed-8945-11e459e54f25" containerName="dnsmasq-dns" containerID="cri-o://621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1" gracePeriod=10 Dec 03 07:11:13 crc kubenswrapper[4946]: I1203 07:11:13.933413 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:13 crc kubenswrapper[4946]: I1203 07:11:13.965507 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" podStartSLOduration=7.96549172 podStartE2EDuration="7.96549172s" podCreationTimestamp="2025-12-03 07:11:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:13.958019531 +0000 UTC m=+1266.754709650" watchObservedRunningTime="2025-12-03 07:11:13.96549172 +0000 UTC m=+1266.762181829" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.816410 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.881498 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgpnz\" (UniqueName: \"kubernetes.io/projected/c33d80bb-9824-45ed-8945-11e459e54f25-kube-api-access-fgpnz\") pod \"c33d80bb-9824-45ed-8945-11e459e54f25\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.881850 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-nb\") pod \"c33d80bb-9824-45ed-8945-11e459e54f25\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.881896 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-sb\") pod \"c33d80bb-9824-45ed-8945-11e459e54f25\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.882108 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-config\") pod \"c33d80bb-9824-45ed-8945-11e459e54f25\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.882178 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-dns-svc\") pod \"c33d80bb-9824-45ed-8945-11e459e54f25\" (UID: \"c33d80bb-9824-45ed-8945-11e459e54f25\") " Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.885301 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c33d80bb-9824-45ed-8945-11e459e54f25-kube-api-access-fgpnz" (OuterVolumeSpecName: "kube-api-access-fgpnz") pod "c33d80bb-9824-45ed-8945-11e459e54f25" (UID: "c33d80bb-9824-45ed-8945-11e459e54f25"). InnerVolumeSpecName "kube-api-access-fgpnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.920937 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c33d80bb-9824-45ed-8945-11e459e54f25" (UID: "c33d80bb-9824-45ed-8945-11e459e54f25"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.926870 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c33d80bb-9824-45ed-8945-11e459e54f25" (UID: "c33d80bb-9824-45ed-8945-11e459e54f25"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.928398 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c33d80bb-9824-45ed-8945-11e459e54f25" (UID: "c33d80bb-9824-45ed-8945-11e459e54f25"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.936836 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-config" (OuterVolumeSpecName: "config") pod "c33d80bb-9824-45ed-8945-11e459e54f25" (UID: "c33d80bb-9824-45ed-8945-11e459e54f25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.941299 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"80557a83-debb-4e3a-a2d0-bb5ac72d824c","Type":"ContainerStarted","Data":"1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d"} Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.943219 4946 generic.go:334] "Generic (PLEG): container finished" podID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerID="0a71a58692dde04b294a34938de02f2640847f1901fd5f7af9a59255f404836e" exitCode=0 Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.943283 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784d65c867-79wht" event={"ID":"48843445-ca95-4a91-bf4b-05c16ebc945c","Type":"ContainerDied","Data":"0a71a58692dde04b294a34938de02f2640847f1901fd5f7af9a59255f404836e"} Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.946383 4946 generic.go:334] "Generic (PLEG): container finished" podID="c33d80bb-9824-45ed-8945-11e459e54f25" containerID="621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1" exitCode=0 Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.946420 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" event={"ID":"c33d80bb-9824-45ed-8945-11e459e54f25","Type":"ContainerDied","Data":"621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1"} Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.946441 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.946455 4946 scope.go:117] "RemoveContainer" containerID="621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1" Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.946444 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-984c76dd7-rp5sp" event={"ID":"c33d80bb-9824-45ed-8945-11e459e54f25","Type":"ContainerDied","Data":"2f1fb8bfc48d0520ba0c7797fe63271efe3cd7e05503f6e90b0634031e1b0d4d"} Dec 03 07:11:14 crc kubenswrapper[4946]: I1203 07:11:14.975057 4946 scope.go:117] "RemoveContainer" containerID="e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.029830 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-984c76dd7-rp5sp"] Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.039930 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.039959 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.039972 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgpnz\" (UniqueName: \"kubernetes.io/projected/c33d80bb-9824-45ed-8945-11e459e54f25-kube-api-access-fgpnz\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.039983 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.039993 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33d80bb-9824-45ed-8945-11e459e54f25-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.045155 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-984c76dd7-rp5sp"] Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.047297 4946 scope.go:117] "RemoveContainer" containerID="621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1" Dec 03 07:11:15 crc kubenswrapper[4946]: E1203 07:11:15.066998 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1\": container with ID starting with 621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1 not found: ID does not exist" containerID="621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.067051 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1"} err="failed to get container status \"621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1\": rpc error: code = NotFound desc = could not find container \"621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1\": container with ID starting with 621c72ae4cb0e12ff23a396a28cd80905fe422ff4aacaf8467efe257424109b1 not found: ID does not exist" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.067075 4946 scope.go:117] "RemoveContainer" containerID="e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685" Dec 03 07:11:15 crc kubenswrapper[4946]: E1203 07:11:15.071052 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685\": container with ID starting with e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685 not found: ID does not exist" containerID="e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.071090 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685"} err="failed to get container status \"e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685\": rpc error: code = NotFound desc = could not find container \"e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685\": container with ID starting with e3046934d29df9cfb95321f1f2a46e909a082f84660a2295528eba8ea343a685 not found: ID does not exist" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.498972 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:15 crc kubenswrapper[4946]: E1203 07:11:15.499114 4946 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 07:11:15 crc kubenswrapper[4946]: E1203 07:11:15.499628 4946 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 07:11:15 crc kubenswrapper[4946]: E1203 07:11:15.499697 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift podName:54cbc35f-afb5-4901-a64f-5e08fee6fd22 nodeName:}" failed. No retries permitted until 2025-12-03 07:11:19.499678652 +0000 UTC m=+1272.296368761 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift") pod "swift-storage-0" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22") : configmap "swift-ring-files" not found Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.608681 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c33d80bb-9824-45ed-8945-11e459e54f25" path="/var/lib/kubelet/pods/c33d80bb-9824-45ed-8945-11e459e54f25/volumes" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.959971 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"80557a83-debb-4e3a-a2d0-bb5ac72d824c","Type":"ContainerStarted","Data":"6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82"} Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.961901 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.968366 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784d65c867-79wht" event={"ID":"48843445-ca95-4a91-bf4b-05c16ebc945c","Type":"ContainerStarted","Data":"739eb262b67e2f24bd3529b30f8a2ef7d1f0674089c1cd2b2752b1e3e063ec1e"} Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.968649 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.976767 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.247491982 podStartE2EDuration="8.976752225s" podCreationTimestamp="2025-12-03 07:11:07 +0000 UTC" firstStartedPulling="2025-12-03 07:11:07.872918725 +0000 UTC m=+1260.669608834" lastFinishedPulling="2025-12-03 07:11:14.602178968 +0000 UTC m=+1267.398869077" observedRunningTime="2025-12-03 07:11:15.976442597 +0000 UTC m=+1268.773132726" watchObservedRunningTime="2025-12-03 07:11:15.976752225 +0000 UTC m=+1268.773442354" Dec 03 07:11:15 crc kubenswrapper[4946]: I1203 07:11:15.997491 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-784d65c867-79wht" podStartSLOduration=5.997475598 podStartE2EDuration="5.997475598s" podCreationTimestamp="2025-12-03 07:11:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:15.991582771 +0000 UTC m=+1268.788272880" watchObservedRunningTime="2025-12-03 07:11:15.997475598 +0000 UTC m=+1268.794165707" Dec 03 07:11:16 crc kubenswrapper[4946]: I1203 07:11:16.124369 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 07:11:16 crc kubenswrapper[4946]: I1203 07:11:16.201416 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 07:11:18 crc kubenswrapper[4946]: I1203 07:11:18.619155 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 07:11:18 crc kubenswrapper[4946]: I1203 07:11:18.737672 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 07:11:19 crc kubenswrapper[4946]: I1203 07:11:19.004477 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-s624s" event={"ID":"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a","Type":"ContainerStarted","Data":"d346ebb93b5866ea909e69f58ddb68b815c76d2379fcd0a1e7f521aa60a3c3fb"} Dec 03 07:11:19 crc kubenswrapper[4946]: I1203 07:11:19.030911 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-s624s" podStartSLOduration=2.173524248 podStartE2EDuration="7.030889058s" podCreationTimestamp="2025-12-03 07:11:12 +0000 UTC" firstStartedPulling="2025-12-03 07:11:13.130802054 +0000 UTC m=+1265.927492163" lastFinishedPulling="2025-12-03 07:11:17.988166854 +0000 UTC m=+1270.784856973" observedRunningTime="2025-12-03 07:11:19.019396321 +0000 UTC m=+1271.816086480" watchObservedRunningTime="2025-12-03 07:11:19.030889058 +0000 UTC m=+1271.827579177" Dec 03 07:11:19 crc kubenswrapper[4946]: I1203 07:11:19.570824 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:19 crc kubenswrapper[4946]: E1203 07:11:19.570990 4946 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 07:11:19 crc kubenswrapper[4946]: E1203 07:11:19.571213 4946 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 07:11:19 crc kubenswrapper[4946]: E1203 07:11:19.571263 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift podName:54cbc35f-afb5-4901-a64f-5e08fee6fd22 nodeName:}" failed. No retries permitted until 2025-12-03 07:11:27.571247751 +0000 UTC m=+1280.367937860 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift") pod "swift-storage-0" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22") : configmap "swift-ring-files" not found Dec 03 07:11:20 crc kubenswrapper[4946]: I1203 07:11:20.871020 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:20 crc kubenswrapper[4946]: I1203 07:11:20.963201 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-jjf2z"] Dec 03 07:11:20 crc kubenswrapper[4946]: I1203 07:11:20.963522 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" podUID="4ba0e22f-5c32-4638-8347-013fa968e479" containerName="dnsmasq-dns" containerID="cri-o://162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5" gracePeriod=10 Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.503829 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.605840 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5j2t\" (UniqueName: \"kubernetes.io/projected/4ba0e22f-5c32-4638-8347-013fa968e479-kube-api-access-v5j2t\") pod \"4ba0e22f-5c32-4638-8347-013fa968e479\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.605970 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-dns-svc\") pod \"4ba0e22f-5c32-4638-8347-013fa968e479\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.606108 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-config\") pod \"4ba0e22f-5c32-4638-8347-013fa968e479\" (UID: \"4ba0e22f-5c32-4638-8347-013fa968e479\") " Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.611753 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba0e22f-5c32-4638-8347-013fa968e479-kube-api-access-v5j2t" (OuterVolumeSpecName: "kube-api-access-v5j2t") pod "4ba0e22f-5c32-4638-8347-013fa968e479" (UID: "4ba0e22f-5c32-4638-8347-013fa968e479"). InnerVolumeSpecName "kube-api-access-v5j2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.655913 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-config" (OuterVolumeSpecName: "config") pod "4ba0e22f-5c32-4638-8347-013fa968e479" (UID: "4ba0e22f-5c32-4638-8347-013fa968e479"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.669714 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ba0e22f-5c32-4638-8347-013fa968e479" (UID: "4ba0e22f-5c32-4638-8347-013fa968e479"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.707626 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.707652 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba0e22f-5c32-4638-8347-013fa968e479-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:21 crc kubenswrapper[4946]: I1203 07:11:21.707662 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5j2t\" (UniqueName: \"kubernetes.io/projected/4ba0e22f-5c32-4638-8347-013fa968e479-kube-api-access-v5j2t\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.039397 4946 generic.go:334] "Generic (PLEG): container finished" podID="4ba0e22f-5c32-4638-8347-013fa968e479" containerID="162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5" exitCode=0 Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.039463 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" event={"ID":"4ba0e22f-5c32-4638-8347-013fa968e479","Type":"ContainerDied","Data":"162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5"} Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.039508 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" event={"ID":"4ba0e22f-5c32-4638-8347-013fa968e479","Type":"ContainerDied","Data":"a5de8e034bcc3aed831766d620d3865eafa93abda2d9701456283decb0082c29"} Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.039524 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-jjf2z" Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.039546 4946 scope.go:117] "RemoveContainer" containerID="162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5" Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.091120 4946 scope.go:117] "RemoveContainer" containerID="13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69" Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.094934 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-jjf2z"] Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.099465 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-jjf2z"] Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.129350 4946 scope.go:117] "RemoveContainer" containerID="162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5" Dec 03 07:11:22 crc kubenswrapper[4946]: E1203 07:11:22.129948 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5\": container with ID starting with 162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5 not found: ID does not exist" containerID="162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5" Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.129994 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5"} err="failed to get container status \"162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5\": rpc error: code = NotFound desc = could not find container \"162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5\": container with ID starting with 162b2a9cff5f5e37f4bbb5416806fb15c907ad11b80f31fde66b1bef242197f5 not found: ID does not exist" Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.130021 4946 scope.go:117] "RemoveContainer" containerID="13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69" Dec 03 07:11:22 crc kubenswrapper[4946]: E1203 07:11:22.130550 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69\": container with ID starting with 13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69 not found: ID does not exist" containerID="13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69" Dec 03 07:11:22 crc kubenswrapper[4946]: I1203 07:11:22.130627 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69"} err="failed to get container status \"13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69\": rpc error: code = NotFound desc = could not find container \"13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69\": container with ID starting with 13ae33bfa70e51ddc4e405cfa8abe17a1d89478d4bcd229a5fa329ce244aaf69 not found: ID does not exist" Dec 03 07:11:23 crc kubenswrapper[4946]: I1203 07:11:23.607472 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba0e22f-5c32-4638-8347-013fa968e479" path="/var/lib/kubelet/pods/4ba0e22f-5c32-4638-8347-013fa968e479/volumes" Dec 03 07:11:25 crc kubenswrapper[4946]: I1203 07:11:25.076874 4946 generic.go:334] "Generic (PLEG): container finished" podID="5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" containerID="d346ebb93b5866ea909e69f58ddb68b815c76d2379fcd0a1e7f521aa60a3c3fb" exitCode=0 Dec 03 07:11:25 crc kubenswrapper[4946]: I1203 07:11:25.076954 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-s624s" event={"ID":"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a","Type":"ContainerDied","Data":"d346ebb93b5866ea909e69f58ddb68b815c76d2379fcd0a1e7f521aa60a3c3fb"} Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.528484 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.610386 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-swiftconf\") pod \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.610473 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-combined-ca-bundle\") pod \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.610519 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-dispersionconf\") pod \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.610551 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkpbx\" (UniqueName: \"kubernetes.io/projected/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-kube-api-access-tkpbx\") pod \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.610646 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-scripts\") pod \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.610716 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-ring-data-devices\") pod \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.610776 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-etc-swift\") pod \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\" (UID: \"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a\") " Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.611478 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" (UID: "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.611989 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" (UID: "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.616236 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-kube-api-access-tkpbx" (OuterVolumeSpecName: "kube-api-access-tkpbx") pod "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" (UID: "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a"). InnerVolumeSpecName "kube-api-access-tkpbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.622033 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" (UID: "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.635203 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-scripts" (OuterVolumeSpecName: "scripts") pod "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" (UID: "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.637607 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" (UID: "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.641554 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" (UID: "5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.712409 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.712442 4946 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.712452 4946 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.712461 4946 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.712470 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.712477 4946 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:26 crc kubenswrapper[4946]: I1203 07:11:26.712487 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkpbx\" (UniqueName: \"kubernetes.io/projected/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a-kube-api-access-tkpbx\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.101467 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-s624s" event={"ID":"5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a","Type":"ContainerDied","Data":"d95fa87979592cadaff668dca31906babc196cee1f928e7e0dd311313888e9b1"} Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.101527 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d95fa87979592cadaff668dca31906babc196cee1f928e7e0dd311313888e9b1" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.101572 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-s624s" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.493776 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.628342 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.633456 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"swift-storage-0\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " pod="openstack/swift-storage-0" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.868163 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.947007 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8fdd-account-create-update-k6bfv"] Dec 03 07:11:27 crc kubenswrapper[4946]: E1203 07:11:27.947652 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c33d80bb-9824-45ed-8945-11e459e54f25" containerName="init" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.947670 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c33d80bb-9824-45ed-8945-11e459e54f25" containerName="init" Dec 03 07:11:27 crc kubenswrapper[4946]: E1203 07:11:27.947691 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba0e22f-5c32-4638-8347-013fa968e479" containerName="init" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.947699 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba0e22f-5c32-4638-8347-013fa968e479" containerName="init" Dec 03 07:11:27 crc kubenswrapper[4946]: E1203 07:11:27.947712 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" containerName="swift-ring-rebalance" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.947722 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" containerName="swift-ring-rebalance" Dec 03 07:11:27 crc kubenswrapper[4946]: E1203 07:11:27.947755 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba0e22f-5c32-4638-8347-013fa968e479" containerName="dnsmasq-dns" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.947763 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba0e22f-5c32-4638-8347-013fa968e479" containerName="dnsmasq-dns" Dec 03 07:11:27 crc kubenswrapper[4946]: E1203 07:11:27.947794 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c33d80bb-9824-45ed-8945-11e459e54f25" containerName="dnsmasq-dns" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.947802 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c33d80bb-9824-45ed-8945-11e459e54f25" containerName="dnsmasq-dns" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.947973 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c33d80bb-9824-45ed-8945-11e459e54f25" containerName="dnsmasq-dns" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.947990 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" containerName="swift-ring-rebalance" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.948008 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba0e22f-5c32-4638-8347-013fa968e479" containerName="dnsmasq-dns" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.948586 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.951352 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 07:11:27 crc kubenswrapper[4946]: I1203 07:11:27.960267 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8fdd-account-create-update-k6bfv"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.010339 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ndzlb"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.011766 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.020992 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ndzlb"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.040556 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjhw6\" (UniqueName: \"kubernetes.io/projected/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-kube-api-access-zjhw6\") pod \"keystone-8fdd-account-create-update-k6bfv\" (UID: \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\") " pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.040611 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-operator-scripts\") pod \"keystone-8fdd-account-create-update-k6bfv\" (UID: \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\") " pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.141841 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjhw6\" (UniqueName: \"kubernetes.io/projected/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-kube-api-access-zjhw6\") pod \"keystone-8fdd-account-create-update-k6bfv\" (UID: \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\") " pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.141879 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-operator-scripts\") pod \"keystone-8fdd-account-create-update-k6bfv\" (UID: \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\") " pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.141958 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsgzf\" (UniqueName: \"kubernetes.io/projected/dff5e6e2-1e97-4aef-8cde-be3c301cb763-kube-api-access-jsgzf\") pod \"keystone-db-create-ndzlb\" (UID: \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\") " pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.142013 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dff5e6e2-1e97-4aef-8cde-be3c301cb763-operator-scripts\") pod \"keystone-db-create-ndzlb\" (UID: \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\") " pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.143101 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-operator-scripts\") pod \"keystone-8fdd-account-create-update-k6bfv\" (UID: \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\") " pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.158515 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjhw6\" (UniqueName: \"kubernetes.io/projected/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-kube-api-access-zjhw6\") pod \"keystone-8fdd-account-create-update-k6bfv\" (UID: \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\") " pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.198059 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-qmrwz"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.199312 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.217568 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-qmrwz"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.243261 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dff5e6e2-1e97-4aef-8cde-be3c301cb763-operator-scripts\") pod \"keystone-db-create-ndzlb\" (UID: \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\") " pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.243336 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17d6a59c-9fc5-47c3-b296-a32eec387ade-operator-scripts\") pod \"placement-db-create-qmrwz\" (UID: \"17d6a59c-9fc5-47c3-b296-a32eec387ade\") " pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.243379 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bff4\" (UniqueName: \"kubernetes.io/projected/17d6a59c-9fc5-47c3-b296-a32eec387ade-kube-api-access-5bff4\") pod \"placement-db-create-qmrwz\" (UID: \"17d6a59c-9fc5-47c3-b296-a32eec387ade\") " pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.243436 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsgzf\" (UniqueName: \"kubernetes.io/projected/dff5e6e2-1e97-4aef-8cde-be3c301cb763-kube-api-access-jsgzf\") pod \"keystone-db-create-ndzlb\" (UID: \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\") " pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.244109 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dff5e6e2-1e97-4aef-8cde-be3c301cb763-operator-scripts\") pod \"keystone-db-create-ndzlb\" (UID: \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\") " pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.258678 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsgzf\" (UniqueName: \"kubernetes.io/projected/dff5e6e2-1e97-4aef-8cde-be3c301cb763-kube-api-access-jsgzf\") pod \"keystone-db-create-ndzlb\" (UID: \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\") " pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.300608 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-0960-account-create-update-2xwzs"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.301865 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.303829 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.309107 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0960-account-create-update-2xwzs"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.318134 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.334418 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.352254 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17d6a59c-9fc5-47c3-b296-a32eec387ade-operator-scripts\") pod \"placement-db-create-qmrwz\" (UID: \"17d6a59c-9fc5-47c3-b296-a32eec387ade\") " pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.352330 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bff4\" (UniqueName: \"kubernetes.io/projected/17d6a59c-9fc5-47c3-b296-a32eec387ade-kube-api-access-5bff4\") pod \"placement-db-create-qmrwz\" (UID: \"17d6a59c-9fc5-47c3-b296-a32eec387ade\") " pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.352394 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eeb90cf-935a-4792-a862-31ba108d49c8-operator-scripts\") pod \"placement-0960-account-create-update-2xwzs\" (UID: \"9eeb90cf-935a-4792-a862-31ba108d49c8\") " pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.352521 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78nhm\" (UniqueName: \"kubernetes.io/projected/9eeb90cf-935a-4792-a862-31ba108d49c8-kube-api-access-78nhm\") pod \"placement-0960-account-create-update-2xwzs\" (UID: \"9eeb90cf-935a-4792-a862-31ba108d49c8\") " pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.353321 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17d6a59c-9fc5-47c3-b296-a32eec387ade-operator-scripts\") pod \"placement-db-create-qmrwz\" (UID: \"17d6a59c-9fc5-47c3-b296-a32eec387ade\") " pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.373886 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bff4\" (UniqueName: \"kubernetes.io/projected/17d6a59c-9fc5-47c3-b296-a32eec387ade-kube-api-access-5bff4\") pod \"placement-db-create-qmrwz\" (UID: \"17d6a59c-9fc5-47c3-b296-a32eec387ade\") " pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.458464 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.459433 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eeb90cf-935a-4792-a862-31ba108d49c8-operator-scripts\") pod \"placement-0960-account-create-update-2xwzs\" (UID: \"9eeb90cf-935a-4792-a862-31ba108d49c8\") " pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.459544 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78nhm\" (UniqueName: \"kubernetes.io/projected/9eeb90cf-935a-4792-a862-31ba108d49c8-kube-api-access-78nhm\") pod \"placement-0960-account-create-update-2xwzs\" (UID: \"9eeb90cf-935a-4792-a862-31ba108d49c8\") " pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.460706 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eeb90cf-935a-4792-a862-31ba108d49c8-operator-scripts\") pod \"placement-0960-account-create-update-2xwzs\" (UID: \"9eeb90cf-935a-4792-a862-31ba108d49c8\") " pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.489760 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78nhm\" (UniqueName: \"kubernetes.io/projected/9eeb90cf-935a-4792-a862-31ba108d49c8-kube-api-access-78nhm\") pod \"placement-0960-account-create-update-2xwzs\" (UID: \"9eeb90cf-935a-4792-a862-31ba108d49c8\") " pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.501559 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8l256"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.503389 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8l256" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.515547 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.519979 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c521-account-create-update-48cv5"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.521267 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.523313 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.537394 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8l256"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.545325 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c521-account-create-update-48cv5"] Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.560688 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c6a0a88-de49-4843-b241-1c0d25c6230a-operator-scripts\") pod \"glance-c521-account-create-update-48cv5\" (UID: \"3c6a0a88-de49-4843-b241-1c0d25c6230a\") " pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.560728 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9572139a-335a-417b-a498-66408e40ae18-operator-scripts\") pod \"glance-db-create-8l256\" (UID: \"9572139a-335a-417b-a498-66408e40ae18\") " pod="openstack/glance-db-create-8l256" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.560761 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkzjq\" (UniqueName: \"kubernetes.io/projected/9572139a-335a-417b-a498-66408e40ae18-kube-api-access-wkzjq\") pod \"glance-db-create-8l256\" (UID: \"9572139a-335a-417b-a498-66408e40ae18\") " pod="openstack/glance-db-create-8l256" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.560871 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wx6m\" (UniqueName: \"kubernetes.io/projected/3c6a0a88-de49-4843-b241-1c0d25c6230a-kube-api-access-8wx6m\") pod \"glance-c521-account-create-update-48cv5\" (UID: \"3c6a0a88-de49-4843-b241-1c0d25c6230a\") " pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.627440 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.665499 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wx6m\" (UniqueName: \"kubernetes.io/projected/3c6a0a88-de49-4843-b241-1c0d25c6230a-kube-api-access-8wx6m\") pod \"glance-c521-account-create-update-48cv5\" (UID: \"3c6a0a88-de49-4843-b241-1c0d25c6230a\") " pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.665916 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c6a0a88-de49-4843-b241-1c0d25c6230a-operator-scripts\") pod \"glance-c521-account-create-update-48cv5\" (UID: \"3c6a0a88-de49-4843-b241-1c0d25c6230a\") " pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.665939 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9572139a-335a-417b-a498-66408e40ae18-operator-scripts\") pod \"glance-db-create-8l256\" (UID: \"9572139a-335a-417b-a498-66408e40ae18\") " pod="openstack/glance-db-create-8l256" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.665953 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkzjq\" (UniqueName: \"kubernetes.io/projected/9572139a-335a-417b-a498-66408e40ae18-kube-api-access-wkzjq\") pod \"glance-db-create-8l256\" (UID: \"9572139a-335a-417b-a498-66408e40ae18\") " pod="openstack/glance-db-create-8l256" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.668101 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c6a0a88-de49-4843-b241-1c0d25c6230a-operator-scripts\") pod \"glance-c521-account-create-update-48cv5\" (UID: \"3c6a0a88-de49-4843-b241-1c0d25c6230a\") " pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.668456 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9572139a-335a-417b-a498-66408e40ae18-operator-scripts\") pod \"glance-db-create-8l256\" (UID: \"9572139a-335a-417b-a498-66408e40ae18\") " pod="openstack/glance-db-create-8l256" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.686031 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkzjq\" (UniqueName: \"kubernetes.io/projected/9572139a-335a-417b-a498-66408e40ae18-kube-api-access-wkzjq\") pod \"glance-db-create-8l256\" (UID: \"9572139a-335a-417b-a498-66408e40ae18\") " pod="openstack/glance-db-create-8l256" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.688658 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wx6m\" (UniqueName: \"kubernetes.io/projected/3c6a0a88-de49-4843-b241-1c0d25c6230a-kube-api-access-8wx6m\") pod \"glance-c521-account-create-update-48cv5\" (UID: \"3c6a0a88-de49-4843-b241-1c0d25c6230a\") " pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.782053 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8fdd-account-create-update-k6bfv"] Dec 03 07:11:28 crc kubenswrapper[4946]: W1203 07:11:28.785593 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e8e38b6_90cf_4c68_ac4e_06b2ee1e30cd.slice/crio-77c280c2690da9172f2bd254dff92c00a184ff90be2301dce1c659c6e5810b6e WatchSource:0}: Error finding container 77c280c2690da9172f2bd254dff92c00a184ff90be2301dce1c659c6e5810b6e: Status 404 returned error can't find the container with id 77c280c2690da9172f2bd254dff92c00a184ff90be2301dce1c659c6e5810b6e Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.823551 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8l256" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.841461 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.847875 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ndzlb"] Dec 03 07:11:28 crc kubenswrapper[4946]: W1203 07:11:28.872109 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddff5e6e2_1e97_4aef_8cde_be3c301cb763.slice/crio-6fac46f5a0bfa8f2394e8ffa19c5d8c0da04adc6769f323b90dd1288ce56a780 WatchSource:0}: Error finding container 6fac46f5a0bfa8f2394e8ffa19c5d8c0da04adc6769f323b90dd1288ce56a780: Status 404 returned error can't find the container with id 6fac46f5a0bfa8f2394e8ffa19c5d8c0da04adc6769f323b90dd1288ce56a780 Dec 03 07:11:28 crc kubenswrapper[4946]: I1203 07:11:28.990694 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-qmrwz"] Dec 03 07:11:29 crc kubenswrapper[4946]: W1203 07:11:29.004398 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17d6a59c_9fc5_47c3_b296_a32eec387ade.slice/crio-576652da43577d075fdfc1d130d7cc9104ee8a806699a3eadeacf943e71d261d WatchSource:0}: Error finding container 576652da43577d075fdfc1d130d7cc9104ee8a806699a3eadeacf943e71d261d: Status 404 returned error can't find the container with id 576652da43577d075fdfc1d130d7cc9104ee8a806699a3eadeacf943e71d261d Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.102828 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0960-account-create-update-2xwzs"] Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.154047 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"1793d3ed6844ec1e05f8d0259de7655d2a4b57adf23a1c3084e2c26c023d072d"} Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.162014 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ndzlb" event={"ID":"dff5e6e2-1e97-4aef-8cde-be3c301cb763","Type":"ContainerStarted","Data":"d697ca9b1e80be206a399092b24806b7f6fa0b97b98856c8b392319cdac0c1d4"} Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.162058 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ndzlb" event={"ID":"dff5e6e2-1e97-4aef-8cde-be3c301cb763","Type":"ContainerStarted","Data":"6fac46f5a0bfa8f2394e8ffa19c5d8c0da04adc6769f323b90dd1288ce56a780"} Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.163271 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qmrwz" event={"ID":"17d6a59c-9fc5-47c3-b296-a32eec387ade","Type":"ContainerStarted","Data":"576652da43577d075fdfc1d130d7cc9104ee8a806699a3eadeacf943e71d261d"} Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.165391 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8fdd-account-create-update-k6bfv" event={"ID":"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd","Type":"ContainerStarted","Data":"6700c2fcb08611e2d7bdcdf960bb02fa23e8b4f20b45767e6216ee45698f3fc5"} Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.165424 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8fdd-account-create-update-k6bfv" event={"ID":"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd","Type":"ContainerStarted","Data":"77c280c2690da9172f2bd254dff92c00a184ff90be2301dce1c659c6e5810b6e"} Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.176346 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-ndzlb" podStartSLOduration=2.176329084 podStartE2EDuration="2.176329084s" podCreationTimestamp="2025-12-03 07:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:29.174296319 +0000 UTC m=+1281.970986428" watchObservedRunningTime="2025-12-03 07:11:29.176329084 +0000 UTC m=+1281.973019193" Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.197351 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-qmrwz" podStartSLOduration=1.197331855 podStartE2EDuration="1.197331855s" podCreationTimestamp="2025-12-03 07:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:29.185020366 +0000 UTC m=+1281.981710465" watchObservedRunningTime="2025-12-03 07:11:29.197331855 +0000 UTC m=+1281.994021964" Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.203214 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8fdd-account-create-update-k6bfv" podStartSLOduration=2.203200371 podStartE2EDuration="2.203200371s" podCreationTimestamp="2025-12-03 07:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:29.1956519 +0000 UTC m=+1281.992342009" watchObservedRunningTime="2025-12-03 07:11:29.203200371 +0000 UTC m=+1281.999890470" Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.262033 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8l256"] Dec 03 07:11:29 crc kubenswrapper[4946]: I1203 07:11:29.371440 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c521-account-create-update-48cv5"] Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.172819 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hc87l" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" probeResult="failure" output=< Dec 03 07:11:30 crc kubenswrapper[4946]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 07:11:30 crc kubenswrapper[4946]: > Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.176451 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c521-account-create-update-48cv5" event={"ID":"3c6a0a88-de49-4843-b241-1c0d25c6230a","Type":"ContainerStarted","Data":"1c6942a4964ff011f829ade85e97de580a3421d534d44522ff6248ad37aa57cb"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.176516 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c521-account-create-update-48cv5" event={"ID":"3c6a0a88-de49-4843-b241-1c0d25c6230a","Type":"ContainerStarted","Data":"c9dac65fe8c8d3f3e6a98a2183fbb21bfe223d298937aa6b20df467681bc5875"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.180248 4946 generic.go:334] "Generic (PLEG): container finished" podID="9eeb90cf-935a-4792-a862-31ba108d49c8" containerID="8d2fc8922be755e466d97af0830a141219adb403d1d1638734857140a464dd7a" exitCode=0 Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.180309 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0960-account-create-update-2xwzs" event={"ID":"9eeb90cf-935a-4792-a862-31ba108d49c8","Type":"ContainerDied","Data":"8d2fc8922be755e466d97af0830a141219adb403d1d1638734857140a464dd7a"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.180330 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0960-account-create-update-2xwzs" event={"ID":"9eeb90cf-935a-4792-a862-31ba108d49c8","Type":"ContainerStarted","Data":"0c7219183d8d870bb90bef5b38b195952e100702c7765f92c79e0a6ecf81754c"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.184433 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.188356 4946 generic.go:334] "Generic (PLEG): container finished" podID="dff5e6e2-1e97-4aef-8cde-be3c301cb763" containerID="d697ca9b1e80be206a399092b24806b7f6fa0b97b98856c8b392319cdac0c1d4" exitCode=0 Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.188428 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ndzlb" event={"ID":"dff5e6e2-1e97-4aef-8cde-be3c301cb763","Type":"ContainerDied","Data":"d697ca9b1e80be206a399092b24806b7f6fa0b97b98856c8b392319cdac0c1d4"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.203198 4946 generic.go:334] "Generic (PLEG): container finished" podID="17d6a59c-9fc5-47c3-b296-a32eec387ade" containerID="a4c5ab9a4b7288722bd3503516f639271ce0628e104cf87e262381e3ac10cc21" exitCode=0 Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.203262 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qmrwz" event={"ID":"17d6a59c-9fc5-47c3-b296-a32eec387ade","Type":"ContainerDied","Data":"a4c5ab9a4b7288722bd3503516f639271ce0628e104cf87e262381e3ac10cc21"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.205555 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8l256" event={"ID":"9572139a-335a-417b-a498-66408e40ae18","Type":"ContainerStarted","Data":"f3336a95769c42e445e00fea4ac613e0e2b8245b491ca7999ea149e6287a6f15"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.205597 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8l256" event={"ID":"9572139a-335a-417b-a498-66408e40ae18","Type":"ContainerStarted","Data":"1c56f7247c79e8519600026c3d82aa68cfdd1aeca02372db238482fed397b20b"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.210601 4946 generic.go:334] "Generic (PLEG): container finished" podID="7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd" containerID="6700c2fcb08611e2d7bdcdf960bb02fa23e8b4f20b45767e6216ee45698f3fc5" exitCode=0 Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.210647 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8fdd-account-create-update-k6bfv" event={"ID":"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd","Type":"ContainerDied","Data":"6700c2fcb08611e2d7bdcdf960bb02fa23e8b4f20b45767e6216ee45698f3fc5"} Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.230593 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-c521-account-create-update-48cv5" podStartSLOduration=2.230574254 podStartE2EDuration="2.230574254s" podCreationTimestamp="2025-12-03 07:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:30.203215723 +0000 UTC m=+1282.999905832" watchObservedRunningTime="2025-12-03 07:11:30.230574254 +0000 UTC m=+1283.027264363" Dec 03 07:11:30 crc kubenswrapper[4946]: I1203 07:11:30.287863 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-8l256" podStartSLOduration=2.287810663 podStartE2EDuration="2.287810663s" podCreationTimestamp="2025-12-03 07:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:30.284061403 +0000 UTC m=+1283.080751522" watchObservedRunningTime="2025-12-03 07:11:30.287810663 +0000 UTC m=+1283.084500792" Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.230001 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a"} Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.230477 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1"} Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.230542 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c"} Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.233137 4946 generic.go:334] "Generic (PLEG): container finished" podID="9572139a-335a-417b-a498-66408e40ae18" containerID="f3336a95769c42e445e00fea4ac613e0e2b8245b491ca7999ea149e6287a6f15" exitCode=0 Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.233288 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8l256" event={"ID":"9572139a-335a-417b-a498-66408e40ae18","Type":"ContainerDied","Data":"f3336a95769c42e445e00fea4ac613e0e2b8245b491ca7999ea149e6287a6f15"} Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.241090 4946 generic.go:334] "Generic (PLEG): container finished" podID="3c6a0a88-de49-4843-b241-1c0d25c6230a" containerID="1c6942a4964ff011f829ade85e97de580a3421d534d44522ff6248ad37aa57cb" exitCode=0 Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.241245 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c521-account-create-update-48cv5" event={"ID":"3c6a0a88-de49-4843-b241-1c0d25c6230a","Type":"ContainerDied","Data":"1c6942a4964ff011f829ade85e97de580a3421d534d44522ff6248ad37aa57cb"} Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.922261 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.931486 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:31 crc kubenswrapper[4946]: I1203 07:11:31.996290 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.019121 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.041904 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dff5e6e2-1e97-4aef-8cde-be3c301cb763-operator-scripts\") pod \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\" (UID: \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.042413 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsgzf\" (UniqueName: \"kubernetes.io/projected/dff5e6e2-1e97-4aef-8cde-be3c301cb763-kube-api-access-jsgzf\") pod \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\" (UID: \"dff5e6e2-1e97-4aef-8cde-be3c301cb763\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.042535 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78nhm\" (UniqueName: \"kubernetes.io/projected/9eeb90cf-935a-4792-a862-31ba108d49c8-kube-api-access-78nhm\") pod \"9eeb90cf-935a-4792-a862-31ba108d49c8\" (UID: \"9eeb90cf-935a-4792-a862-31ba108d49c8\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.042824 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eeb90cf-935a-4792-a862-31ba108d49c8-operator-scripts\") pod \"9eeb90cf-935a-4792-a862-31ba108d49c8\" (UID: \"9eeb90cf-935a-4792-a862-31ba108d49c8\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.042951 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjhw6\" (UniqueName: \"kubernetes.io/projected/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-kube-api-access-zjhw6\") pod \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\" (UID: \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.043087 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-operator-scripts\") pod \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\" (UID: \"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.045664 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dff5e6e2-1e97-4aef-8cde-be3c301cb763-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dff5e6e2-1e97-4aef-8cde-be3c301cb763" (UID: "dff5e6e2-1e97-4aef-8cde-be3c301cb763"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.045809 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eeb90cf-935a-4792-a862-31ba108d49c8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9eeb90cf-935a-4792-a862-31ba108d49c8" (UID: "9eeb90cf-935a-4792-a862-31ba108d49c8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.048748 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd" (UID: "7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.050992 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dff5e6e2-1e97-4aef-8cde-be3c301cb763-kube-api-access-jsgzf" (OuterVolumeSpecName: "kube-api-access-jsgzf") pod "dff5e6e2-1e97-4aef-8cde-be3c301cb763" (UID: "dff5e6e2-1e97-4aef-8cde-be3c301cb763"). InnerVolumeSpecName "kube-api-access-jsgzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.052071 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eeb90cf-935a-4792-a862-31ba108d49c8-kube-api-access-78nhm" (OuterVolumeSpecName: "kube-api-access-78nhm") pod "9eeb90cf-935a-4792-a862-31ba108d49c8" (UID: "9eeb90cf-935a-4792-a862-31ba108d49c8"). InnerVolumeSpecName "kube-api-access-78nhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.053151 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-kube-api-access-zjhw6" (OuterVolumeSpecName: "kube-api-access-zjhw6") pod "7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd" (UID: "7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd"). InnerVolumeSpecName "kube-api-access-zjhw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.147426 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bff4\" (UniqueName: \"kubernetes.io/projected/17d6a59c-9fc5-47c3-b296-a32eec387ade-kube-api-access-5bff4\") pod \"17d6a59c-9fc5-47c3-b296-a32eec387ade\" (UID: \"17d6a59c-9fc5-47c3-b296-a32eec387ade\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.147489 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17d6a59c-9fc5-47c3-b296-a32eec387ade-operator-scripts\") pod \"17d6a59c-9fc5-47c3-b296-a32eec387ade\" (UID: \"17d6a59c-9fc5-47c3-b296-a32eec387ade\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.148305 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eeb90cf-935a-4792-a862-31ba108d49c8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.148326 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjhw6\" (UniqueName: \"kubernetes.io/projected/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-kube-api-access-zjhw6\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.148337 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.148347 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dff5e6e2-1e97-4aef-8cde-be3c301cb763-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.148356 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsgzf\" (UniqueName: \"kubernetes.io/projected/dff5e6e2-1e97-4aef-8cde-be3c301cb763-kube-api-access-jsgzf\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.148365 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78nhm\" (UniqueName: \"kubernetes.io/projected/9eeb90cf-935a-4792-a862-31ba108d49c8-kube-api-access-78nhm\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.148482 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17d6a59c-9fc5-47c3-b296-a32eec387ade-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "17d6a59c-9fc5-47c3-b296-a32eec387ade" (UID: "17d6a59c-9fc5-47c3-b296-a32eec387ade"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.151315 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d6a59c-9fc5-47c3-b296-a32eec387ade-kube-api-access-5bff4" (OuterVolumeSpecName: "kube-api-access-5bff4") pod "17d6a59c-9fc5-47c3-b296-a32eec387ade" (UID: "17d6a59c-9fc5-47c3-b296-a32eec387ade"). InnerVolumeSpecName "kube-api-access-5bff4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.249458 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bff4\" (UniqueName: \"kubernetes.io/projected/17d6a59c-9fc5-47c3-b296-a32eec387ade-kube-api-access-5bff4\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.249484 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17d6a59c-9fc5-47c3-b296-a32eec387ade-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.249819 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0960-account-create-update-2xwzs" event={"ID":"9eeb90cf-935a-4792-a862-31ba108d49c8","Type":"ContainerDied","Data":"0c7219183d8d870bb90bef5b38b195952e100702c7765f92c79e0a6ecf81754c"} Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.249849 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c7219183d8d870bb90bef5b38b195952e100702c7765f92c79e0a6ecf81754c" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.249906 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0960-account-create-update-2xwzs" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.254348 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7"} Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.255842 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ndzlb" event={"ID":"dff5e6e2-1e97-4aef-8cde-be3c301cb763","Type":"ContainerDied","Data":"6fac46f5a0bfa8f2394e8ffa19c5d8c0da04adc6769f323b90dd1288ce56a780"} Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.255885 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fac46f5a0bfa8f2394e8ffa19c5d8c0da04adc6769f323b90dd1288ce56a780" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.256001 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ndzlb" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.260983 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qmrwz" event={"ID":"17d6a59c-9fc5-47c3-b296-a32eec387ade","Type":"ContainerDied","Data":"576652da43577d075fdfc1d130d7cc9104ee8a806699a3eadeacf943e71d261d"} Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.261018 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="576652da43577d075fdfc1d130d7cc9104ee8a806699a3eadeacf943e71d261d" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.261036 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qmrwz" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.263916 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8fdd-account-create-update-k6bfv" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.267642 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8fdd-account-create-update-k6bfv" event={"ID":"7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd","Type":"ContainerDied","Data":"77c280c2690da9172f2bd254dff92c00a184ff90be2301dce1c659c6e5810b6e"} Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.267700 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77c280c2690da9172f2bd254dff92c00a184ff90be2301dce1c659c6e5810b6e" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.569144 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8l256" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.603974 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.655406 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wx6m\" (UniqueName: \"kubernetes.io/projected/3c6a0a88-de49-4843-b241-1c0d25c6230a-kube-api-access-8wx6m\") pod \"3c6a0a88-de49-4843-b241-1c0d25c6230a\" (UID: \"3c6a0a88-de49-4843-b241-1c0d25c6230a\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.655513 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9572139a-335a-417b-a498-66408e40ae18-operator-scripts\") pod \"9572139a-335a-417b-a498-66408e40ae18\" (UID: \"9572139a-335a-417b-a498-66408e40ae18\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.655536 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkzjq\" (UniqueName: \"kubernetes.io/projected/9572139a-335a-417b-a498-66408e40ae18-kube-api-access-wkzjq\") pod \"9572139a-335a-417b-a498-66408e40ae18\" (UID: \"9572139a-335a-417b-a498-66408e40ae18\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.655675 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c6a0a88-de49-4843-b241-1c0d25c6230a-operator-scripts\") pod \"3c6a0a88-de49-4843-b241-1c0d25c6230a\" (UID: \"3c6a0a88-de49-4843-b241-1c0d25c6230a\") " Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.656508 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9572139a-335a-417b-a498-66408e40ae18-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9572139a-335a-417b-a498-66408e40ae18" (UID: "9572139a-335a-417b-a498-66408e40ae18"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.656940 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c6a0a88-de49-4843-b241-1c0d25c6230a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c6a0a88-de49-4843-b241-1c0d25c6230a" (UID: "3c6a0a88-de49-4843-b241-1c0d25c6230a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.656974 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9572139a-335a-417b-a498-66408e40ae18-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.659786 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9572139a-335a-417b-a498-66408e40ae18-kube-api-access-wkzjq" (OuterVolumeSpecName: "kube-api-access-wkzjq") pod "9572139a-335a-417b-a498-66408e40ae18" (UID: "9572139a-335a-417b-a498-66408e40ae18"). InnerVolumeSpecName "kube-api-access-wkzjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.659880 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c6a0a88-de49-4843-b241-1c0d25c6230a-kube-api-access-8wx6m" (OuterVolumeSpecName: "kube-api-access-8wx6m") pod "3c6a0a88-de49-4843-b241-1c0d25c6230a" (UID: "3c6a0a88-de49-4843-b241-1c0d25c6230a"). InnerVolumeSpecName "kube-api-access-8wx6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.757821 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c6a0a88-de49-4843-b241-1c0d25c6230a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.757865 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wx6m\" (UniqueName: \"kubernetes.io/projected/3c6a0a88-de49-4843-b241-1c0d25c6230a-kube-api-access-8wx6m\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:32 crc kubenswrapper[4946]: I1203 07:11:32.757878 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkzjq\" (UniqueName: \"kubernetes.io/projected/9572139a-335a-417b-a498-66408e40ae18-kube-api-access-wkzjq\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.277222 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c521-account-create-update-48cv5" event={"ID":"3c6a0a88-de49-4843-b241-1c0d25c6230a","Type":"ContainerDied","Data":"c9dac65fe8c8d3f3e6a98a2183fbb21bfe223d298937aa6b20df467681bc5875"} Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.277269 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9dac65fe8c8d3f3e6a98a2183fbb21bfe223d298937aa6b20df467681bc5875" Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.277282 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c521-account-create-update-48cv5" Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.282311 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51"} Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.282353 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e"} Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.282363 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2"} Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.284762 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8l256" event={"ID":"9572139a-335a-417b-a498-66408e40ae18","Type":"ContainerDied","Data":"1c56f7247c79e8519600026c3d82aa68cfdd1aeca02372db238482fed397b20b"} Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.284788 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c56f7247c79e8519600026c3d82aa68cfdd1aeca02372db238482fed397b20b" Dec 03 07:11:33 crc kubenswrapper[4946]: I1203 07:11:33.284854 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8l256" Dec 03 07:11:34 crc kubenswrapper[4946]: I1203 07:11:34.295771 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59"} Dec 03 07:11:34 crc kubenswrapper[4946]: I1203 07:11:34.295819 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a"} Dec 03 07:11:34 crc kubenswrapper[4946]: I1203 07:11:34.296935 4946 generic.go:334] "Generic (PLEG): container finished" podID="2f821218-d2a4-4f97-a924-199c056c7c55" containerID="15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d" exitCode=0 Dec 03 07:11:34 crc kubenswrapper[4946]: I1203 07:11:34.296996 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2f821218-d2a4-4f97-a924-199c056c7c55","Type":"ContainerDied","Data":"15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d"} Dec 03 07:11:34 crc kubenswrapper[4946]: I1203 07:11:34.299605 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"854575dc-d77d-43fe-a7ff-495ef76ab456","Type":"ContainerDied","Data":"5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78"} Dec 03 07:11:34 crc kubenswrapper[4946]: I1203 07:11:34.299561 4946 generic.go:334] "Generic (PLEG): container finished" podID="854575dc-d77d-43fe-a7ff-495ef76ab456" containerID="5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78" exitCode=0 Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.163246 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hc87l" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" probeResult="failure" output=< Dec 03 07:11:35 crc kubenswrapper[4946]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 07:11:35 crc kubenswrapper[4946]: > Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.241378 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.256811 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.313660 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"854575dc-d77d-43fe-a7ff-495ef76ab456","Type":"ContainerStarted","Data":"2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32"} Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.313863 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.330595 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a"} Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.330633 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d"} Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.330643 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19"} Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.330651 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973"} Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.332830 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2f821218-d2a4-4f97-a924-199c056c7c55","Type":"ContainerStarted","Data":"7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e"} Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.333184 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.366978 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=55.118190459 podStartE2EDuration="1m2.366961458s" podCreationTimestamp="2025-12-03 07:10:33 +0000 UTC" firstStartedPulling="2025-12-03 07:10:51.775259383 +0000 UTC m=+1244.571949492" lastFinishedPulling="2025-12-03 07:10:59.024030382 +0000 UTC m=+1251.820720491" observedRunningTime="2025-12-03 07:11:35.361299167 +0000 UTC m=+1288.157989276" watchObservedRunningTime="2025-12-03 07:11:35.366961458 +0000 UTC m=+1288.163651567" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.394523 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=55.134886964 podStartE2EDuration="1m2.394509614s" podCreationTimestamp="2025-12-03 07:10:33 +0000 UTC" firstStartedPulling="2025-12-03 07:10:51.322346864 +0000 UTC m=+1244.119036973" lastFinishedPulling="2025-12-03 07:10:58.581969494 +0000 UTC m=+1251.378659623" observedRunningTime="2025-12-03 07:11:35.392630364 +0000 UTC m=+1288.189320473" watchObservedRunningTime="2025-12-03 07:11:35.394509614 +0000 UTC m=+1288.191199723" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468210 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hc87l-config-mk9k2"] Dec 03 07:11:35 crc kubenswrapper[4946]: E1203 07:11:35.468572 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9572139a-335a-417b-a498-66408e40ae18" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468592 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9572139a-335a-417b-a498-66408e40ae18" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: E1203 07:11:35.468612 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468623 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: E1203 07:11:35.468636 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eeb90cf-935a-4792-a862-31ba108d49c8" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468645 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eeb90cf-935a-4792-a862-31ba108d49c8" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: E1203 07:11:35.468662 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d6a59c-9fc5-47c3-b296-a32eec387ade" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468670 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d6a59c-9fc5-47c3-b296-a32eec387ade" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: E1203 07:11:35.468683 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff5e6e2-1e97-4aef-8cde-be3c301cb763" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468692 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff5e6e2-1e97-4aef-8cde-be3c301cb763" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: E1203 07:11:35.468716 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c6a0a88-de49-4843-b241-1c0d25c6230a" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468724 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c6a0a88-de49-4843-b241-1c0d25c6230a" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468928 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="9572139a-335a-417b-a498-66408e40ae18" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468949 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eeb90cf-935a-4792-a862-31ba108d49c8" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468968 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d6a59c-9fc5-47c3-b296-a32eec387ade" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.468990 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.469010 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="dff5e6e2-1e97-4aef-8cde-be3c301cb763" containerName="mariadb-database-create" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.469018 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c6a0a88-de49-4843-b241-1c0d25c6230a" containerName="mariadb-account-create-update" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.469645 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.476206 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.480296 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hc87l-config-mk9k2"] Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.510647 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.510726 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx58g\" (UniqueName: \"kubernetes.io/projected/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-kube-api-access-xx58g\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.510775 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run-ovn\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.510822 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-log-ovn\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.510847 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-scripts\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.510904 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-additional-scripts\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.612606 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.612691 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx58g\" (UniqueName: \"kubernetes.io/projected/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-kube-api-access-xx58g\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.612719 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run-ovn\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.612776 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-log-ovn\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.612799 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-scripts\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.612859 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-additional-scripts\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.613047 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run-ovn\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.613046 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-log-ovn\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.613054 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.613809 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-additional-scripts\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.615032 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-scripts\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.642418 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx58g\" (UniqueName: \"kubernetes.io/projected/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-kube-api-access-xx58g\") pod \"ovn-controller-hc87l-config-mk9k2\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:35 crc kubenswrapper[4946]: I1203 07:11:35.785286 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.283920 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hc87l-config-mk9k2"] Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.345216 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerStarted","Data":"a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499"} Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.346633 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hc87l-config-mk9k2" event={"ID":"f61dffcb-3d3f-47c7-a21b-a4e3e5337407","Type":"ContainerStarted","Data":"d0cb71bc714d5908ba85c9b142141be970079d77a127d26715b66b134d3eb3cc"} Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.694710 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.247371517 podStartE2EDuration="26.694692825s" podCreationTimestamp="2025-12-03 07:11:10 +0000 UTC" firstStartedPulling="2025-12-03 07:11:28.472463622 +0000 UTC m=+1281.269153731" lastFinishedPulling="2025-12-03 07:11:33.91978494 +0000 UTC m=+1286.716475039" observedRunningTime="2025-12-03 07:11:36.394775293 +0000 UTC m=+1289.191465422" watchObservedRunningTime="2025-12-03 07:11:36.694692825 +0000 UTC m=+1289.491382944" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.697201 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d5cc849d9-2r4ds"] Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.698732 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.703230 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.722849 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d5cc849d9-2r4ds"] Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.833290 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-svc\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.833331 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.833354 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.833448 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.833467 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cddqw\" (UniqueName: \"kubernetes.io/projected/af0dc160-2364-4866-b68d-37fe8990241d-kube-api-access-cddqw\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.833498 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-config\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.935071 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.935113 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cddqw\" (UniqueName: \"kubernetes.io/projected/af0dc160-2364-4866-b68d-37fe8990241d-kube-api-access-cddqw\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.935146 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-config\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.935184 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-svc\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.935201 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.935216 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.935995 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.936495 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.937264 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-config\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.937785 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-svc\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.938266 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:36 crc kubenswrapper[4946]: I1203 07:11:36.966725 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cddqw\" (UniqueName: \"kubernetes.io/projected/af0dc160-2364-4866-b68d-37fe8990241d-kube-api-access-cddqw\") pod \"dnsmasq-dns-7d5cc849d9-2r4ds\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:37 crc kubenswrapper[4946]: I1203 07:11:37.098434 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:37 crc kubenswrapper[4946]: W1203 07:11:37.588943 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf0dc160_2364_4866_b68d_37fe8990241d.slice/crio-da198542aabb891a4656828f4f26c609c79ce1f416150dcf2b06bd9e4f517046 WatchSource:0}: Error finding container da198542aabb891a4656828f4f26c609c79ce1f416150dcf2b06bd9e4f517046: Status 404 returned error can't find the container with id da198542aabb891a4656828f4f26c609c79ce1f416150dcf2b06bd9e4f517046 Dec 03 07:11:37 crc kubenswrapper[4946]: I1203 07:11:37.589689 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d5cc849d9-2r4ds"] Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.366418 4946 generic.go:334] "Generic (PLEG): container finished" podID="af0dc160-2364-4866-b68d-37fe8990241d" containerID="75f1b9fda6f1ad88468e26ab6d8942b9d4779f02eff656444fe9648c85ea6137" exitCode=0 Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.366489 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" event={"ID":"af0dc160-2364-4866-b68d-37fe8990241d","Type":"ContainerDied","Data":"75f1b9fda6f1ad88468e26ab6d8942b9d4779f02eff656444fe9648c85ea6137"} Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.366884 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" event={"ID":"af0dc160-2364-4866-b68d-37fe8990241d","Type":"ContainerStarted","Data":"da198542aabb891a4656828f4f26c609c79ce1f416150dcf2b06bd9e4f517046"} Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.370116 4946 generic.go:334] "Generic (PLEG): container finished" podID="f61dffcb-3d3f-47c7-a21b-a4e3e5337407" containerID="2d1d6e59deb3691d32b28325455c408f0366e63c827b8f379c3c4aae9e05ae72" exitCode=0 Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.370138 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hc87l-config-mk9k2" event={"ID":"f61dffcb-3d3f-47c7-a21b-a4e3e5337407","Type":"ContainerDied","Data":"2d1d6e59deb3691d32b28325455c408f0366e63c827b8f379c3c4aae9e05ae72"} Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.583301 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-9q4dr"] Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.584446 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.589520 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.589823 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bk8hz" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.596923 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9q4dr"] Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.663870 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-combined-ca-bundle\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.663931 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-db-sync-config-data\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.664054 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg4pt\" (UniqueName: \"kubernetes.io/projected/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-kube-api-access-pg4pt\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.664110 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-config-data\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.765680 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg4pt\" (UniqueName: \"kubernetes.io/projected/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-kube-api-access-pg4pt\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.765757 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-config-data\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.765821 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-combined-ca-bundle\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.765839 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-db-sync-config-data\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.770192 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-db-sync-config-data\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.770334 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-combined-ca-bundle\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.771051 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-config-data\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.801361 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg4pt\" (UniqueName: \"kubernetes.io/projected/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-kube-api-access-pg4pt\") pod \"glance-db-sync-9q4dr\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:38 crc kubenswrapper[4946]: I1203 07:11:38.967268 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9q4dr" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.384192 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" event={"ID":"af0dc160-2364-4866-b68d-37fe8990241d","Type":"ContainerStarted","Data":"f7fac8d05109eff647ba6132cda7c2e640a242149c274d302f87f99b27b94641"} Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.410158 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" podStartSLOduration=3.41013948 podStartE2EDuration="3.41013948s" podCreationTimestamp="2025-12-03 07:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:39.406188734 +0000 UTC m=+1292.202878853" watchObservedRunningTime="2025-12-03 07:11:39.41013948 +0000 UTC m=+1292.206829589" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.523172 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9q4dr"] Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.705378 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.781821 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-log-ovn\") pod \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.781899 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run\") pod \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.781927 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run-ovn\") pod \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.782001 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-additional-scripts\") pod \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.782074 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-scripts\") pod \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.782154 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx58g\" (UniqueName: \"kubernetes.io/projected/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-kube-api-access-xx58g\") pod \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\" (UID: \"f61dffcb-3d3f-47c7-a21b-a4e3e5337407\") " Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.783518 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f61dffcb-3d3f-47c7-a21b-a4e3e5337407" (UID: "f61dffcb-3d3f-47c7-a21b-a4e3e5337407"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.783558 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f61dffcb-3d3f-47c7-a21b-a4e3e5337407" (UID: "f61dffcb-3d3f-47c7-a21b-a4e3e5337407"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.783575 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run" (OuterVolumeSpecName: "var-run") pod "f61dffcb-3d3f-47c7-a21b-a4e3e5337407" (UID: "f61dffcb-3d3f-47c7-a21b-a4e3e5337407"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.784094 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f61dffcb-3d3f-47c7-a21b-a4e3e5337407" (UID: "f61dffcb-3d3f-47c7-a21b-a4e3e5337407"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.784319 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-scripts" (OuterVolumeSpecName: "scripts") pod "f61dffcb-3d3f-47c7-a21b-a4e3e5337407" (UID: "f61dffcb-3d3f-47c7-a21b-a4e3e5337407"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.788660 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-kube-api-access-xx58g" (OuterVolumeSpecName: "kube-api-access-xx58g") pod "f61dffcb-3d3f-47c7-a21b-a4e3e5337407" (UID: "f61dffcb-3d3f-47c7-a21b-a4e3e5337407"). InnerVolumeSpecName "kube-api-access-xx58g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.884351 4946 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.884386 4946 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.884400 4946 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.884415 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.884441 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx58g\" (UniqueName: \"kubernetes.io/projected/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-kube-api-access-xx58g\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:39 crc kubenswrapper[4946]: I1203 07:11:39.884454 4946 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f61dffcb-3d3f-47c7-a21b-a4e3e5337407-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:40 crc kubenswrapper[4946]: I1203 07:11:40.173235 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-hc87l" Dec 03 07:11:40 crc kubenswrapper[4946]: I1203 07:11:40.392558 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9q4dr" event={"ID":"a59904e6-2de4-4bc6-b0b5-6f578470b7d6","Type":"ContainerStarted","Data":"45f1af885a6ef72f093d1d6772635efcacf75ddbaf5edba61bbd007f700eba72"} Dec 03 07:11:40 crc kubenswrapper[4946]: I1203 07:11:40.396061 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hc87l-config-mk9k2" Dec 03 07:11:40 crc kubenswrapper[4946]: I1203 07:11:40.399945 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hc87l-config-mk9k2" event={"ID":"f61dffcb-3d3f-47c7-a21b-a4e3e5337407","Type":"ContainerDied","Data":"d0cb71bc714d5908ba85c9b142141be970079d77a127d26715b66b134d3eb3cc"} Dec 03 07:11:40 crc kubenswrapper[4946]: I1203 07:11:40.400016 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0cb71bc714d5908ba85c9b142141be970079d77a127d26715b66b134d3eb3cc" Dec 03 07:11:40 crc kubenswrapper[4946]: I1203 07:11:40.400042 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:40 crc kubenswrapper[4946]: I1203 07:11:40.812992 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hc87l-config-mk9k2"] Dec 03 07:11:40 crc kubenswrapper[4946]: I1203 07:11:40.819327 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-hc87l-config-mk9k2"] Dec 03 07:11:41 crc kubenswrapper[4946]: I1203 07:11:41.605338 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f61dffcb-3d3f-47c7-a21b-a4e3e5337407" path="/var/lib/kubelet/pods/f61dffcb-3d3f-47c7-a21b-a4e3e5337407/volumes" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.232548 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.327874 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.532054 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-xj4g5"] Dec 03 07:11:45 crc kubenswrapper[4946]: E1203 07:11:45.532405 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f61dffcb-3d3f-47c7-a21b-a4e3e5337407" containerName="ovn-config" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.532423 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f61dffcb-3d3f-47c7-a21b-a4e3e5337407" containerName="ovn-config" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.532586 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f61dffcb-3d3f-47c7-a21b-a4e3e5337407" containerName="ovn-config" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.533113 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.554270 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xj4g5"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.605136 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-operator-scripts\") pod \"cinder-db-create-xj4g5\" (UID: \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\") " pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.605207 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpx8v\" (UniqueName: \"kubernetes.io/projected/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-kube-api-access-xpx8v\") pod \"cinder-db-create-xj4g5\" (UID: \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\") " pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.631785 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-fftdj"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.632767 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.648347 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-eae6-account-create-update-xrhjp"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.649444 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.652592 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.661360 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-fftdj"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.670784 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-eae6-account-create-update-xrhjp"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.706579 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpx8v\" (UniqueName: \"kubernetes.io/projected/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-kube-api-access-xpx8v\") pod \"cinder-db-create-xj4g5\" (UID: \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\") " pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.706686 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99tjh\" (UniqueName: \"kubernetes.io/projected/cdc5e484-c93b-47bf-8144-e76a5db68bfd-kube-api-access-99tjh\") pod \"barbican-db-create-fftdj\" (UID: \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\") " pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.706726 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880093ae-378a-4b62-902b-35b216d6d022-operator-scripts\") pod \"barbican-eae6-account-create-update-xrhjp\" (UID: \"880093ae-378a-4b62-902b-35b216d6d022\") " pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.706794 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slrls\" (UniqueName: \"kubernetes.io/projected/880093ae-378a-4b62-902b-35b216d6d022-kube-api-access-slrls\") pod \"barbican-eae6-account-create-update-xrhjp\" (UID: \"880093ae-378a-4b62-902b-35b216d6d022\") " pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.706814 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-operator-scripts\") pod \"cinder-db-create-xj4g5\" (UID: \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\") " pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.706848 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc5e484-c93b-47bf-8144-e76a5db68bfd-operator-scripts\") pod \"barbican-db-create-fftdj\" (UID: \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\") " pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.707707 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-operator-scripts\") pod \"cinder-db-create-xj4g5\" (UID: \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\") " pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.727676 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5125-account-create-update-c4zxx"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.728807 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.730889 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.743570 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpx8v\" (UniqueName: \"kubernetes.io/projected/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-kube-api-access-xpx8v\") pod \"cinder-db-create-xj4g5\" (UID: \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\") " pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.747460 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5125-account-create-update-c4zxx"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.808358 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99tjh\" (UniqueName: \"kubernetes.io/projected/cdc5e484-c93b-47bf-8144-e76a5db68bfd-kube-api-access-99tjh\") pod \"barbican-db-create-fftdj\" (UID: \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\") " pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.808413 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880093ae-378a-4b62-902b-35b216d6d022-operator-scripts\") pod \"barbican-eae6-account-create-update-xrhjp\" (UID: \"880093ae-378a-4b62-902b-35b216d6d022\") " pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.808440 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e282365b-8ff7-4fd8-9220-ccdac9865543-operator-scripts\") pod \"cinder-5125-account-create-update-c4zxx\" (UID: \"e282365b-8ff7-4fd8-9220-ccdac9865543\") " pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.808478 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwxp8\" (UniqueName: \"kubernetes.io/projected/e282365b-8ff7-4fd8-9220-ccdac9865543-kube-api-access-wwxp8\") pod \"cinder-5125-account-create-update-c4zxx\" (UID: \"e282365b-8ff7-4fd8-9220-ccdac9865543\") " pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.808507 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slrls\" (UniqueName: \"kubernetes.io/projected/880093ae-378a-4b62-902b-35b216d6d022-kube-api-access-slrls\") pod \"barbican-eae6-account-create-update-xrhjp\" (UID: \"880093ae-378a-4b62-902b-35b216d6d022\") " pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.808533 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc5e484-c93b-47bf-8144-e76a5db68bfd-operator-scripts\") pod \"barbican-db-create-fftdj\" (UID: \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\") " pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.809199 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc5e484-c93b-47bf-8144-e76a5db68bfd-operator-scripts\") pod \"barbican-db-create-fftdj\" (UID: \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\") " pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.809891 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880093ae-378a-4b62-902b-35b216d6d022-operator-scripts\") pod \"barbican-eae6-account-create-update-xrhjp\" (UID: \"880093ae-378a-4b62-902b-35b216d6d022\") " pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.826293 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slrls\" (UniqueName: \"kubernetes.io/projected/880093ae-378a-4b62-902b-35b216d6d022-kube-api-access-slrls\") pod \"barbican-eae6-account-create-update-xrhjp\" (UID: \"880093ae-378a-4b62-902b-35b216d6d022\") " pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.830811 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99tjh\" (UniqueName: \"kubernetes.io/projected/cdc5e484-c93b-47bf-8144-e76a5db68bfd-kube-api-access-99tjh\") pod \"barbican-db-create-fftdj\" (UID: \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\") " pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.851028 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.886834 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-4sm7k"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.887844 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.891022 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.892562 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.892631 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.892712 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mwdjs" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.900333 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4sm7k"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.909493 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e282365b-8ff7-4fd8-9220-ccdac9865543-operator-scripts\") pod \"cinder-5125-account-create-update-c4zxx\" (UID: \"e282365b-8ff7-4fd8-9220-ccdac9865543\") " pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.909758 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwxp8\" (UniqueName: \"kubernetes.io/projected/e282365b-8ff7-4fd8-9220-ccdac9865543-kube-api-access-wwxp8\") pod \"cinder-5125-account-create-update-c4zxx\" (UID: \"e282365b-8ff7-4fd8-9220-ccdac9865543\") " pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.910653 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e282365b-8ff7-4fd8-9220-ccdac9865543-operator-scripts\") pod \"cinder-5125-account-create-update-c4zxx\" (UID: \"e282365b-8ff7-4fd8-9220-ccdac9865543\") " pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.930406 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwxp8\" (UniqueName: \"kubernetes.io/projected/e282365b-8ff7-4fd8-9220-ccdac9865543-kube-api-access-wwxp8\") pod \"cinder-5125-account-create-update-c4zxx\" (UID: \"e282365b-8ff7-4fd8-9220-ccdac9865543\") " pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.952485 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-5f2hw"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.953659 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.959412 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.963850 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6e16-account-create-update-c4t4n"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.964908 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.966142 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.974723 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.979528 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5f2hw"] Dec 03 07:11:45 crc kubenswrapper[4946]: I1203 07:11:45.995617 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6e16-account-create-update-c4t4n"] Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.012610 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-operator-scripts\") pod \"neutron-6e16-account-create-update-c4t4n\" (UID: \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\") " pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.013089 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-config-data\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.013151 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds8dg\" (UniqueName: \"kubernetes.io/projected/4cb069b9-d659-481c-bbb5-616c4af71d65-kube-api-access-ds8dg\") pod \"neutron-db-create-5f2hw\" (UID: \"4cb069b9-d659-481c-bbb5-616c4af71d65\") " pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.013552 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdknx\" (UniqueName: \"kubernetes.io/projected/7a18e489-bc4f-462a-87da-280408bb99e2-kube-api-access-pdknx\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.013591 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb069b9-d659-481c-bbb5-616c4af71d65-operator-scripts\") pod \"neutron-db-create-5f2hw\" (UID: \"4cb069b9-d659-481c-bbb5-616c4af71d65\") " pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.013613 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lk82\" (UniqueName: \"kubernetes.io/projected/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-kube-api-access-8lk82\") pod \"neutron-6e16-account-create-update-c4t4n\" (UID: \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\") " pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.013648 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-combined-ca-bundle\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.083322 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.115016 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-operator-scripts\") pod \"neutron-6e16-account-create-update-c4t4n\" (UID: \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\") " pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.115100 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-config-data\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.115152 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds8dg\" (UniqueName: \"kubernetes.io/projected/4cb069b9-d659-481c-bbb5-616c4af71d65-kube-api-access-ds8dg\") pod \"neutron-db-create-5f2hw\" (UID: \"4cb069b9-d659-481c-bbb5-616c4af71d65\") " pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.115180 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdknx\" (UniqueName: \"kubernetes.io/projected/7a18e489-bc4f-462a-87da-280408bb99e2-kube-api-access-pdknx\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.115198 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb069b9-d659-481c-bbb5-616c4af71d65-operator-scripts\") pod \"neutron-db-create-5f2hw\" (UID: \"4cb069b9-d659-481c-bbb5-616c4af71d65\") " pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.115214 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lk82\" (UniqueName: \"kubernetes.io/projected/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-kube-api-access-8lk82\") pod \"neutron-6e16-account-create-update-c4t4n\" (UID: \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\") " pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.115254 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-combined-ca-bundle\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.116130 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-operator-scripts\") pod \"neutron-6e16-account-create-update-c4t4n\" (UID: \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\") " pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.116802 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb069b9-d659-481c-bbb5-616c4af71d65-operator-scripts\") pod \"neutron-db-create-5f2hw\" (UID: \"4cb069b9-d659-481c-bbb5-616c4af71d65\") " pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.122041 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-config-data\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.133358 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-combined-ca-bundle\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.137860 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lk82\" (UniqueName: \"kubernetes.io/projected/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-kube-api-access-8lk82\") pod \"neutron-6e16-account-create-update-c4t4n\" (UID: \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\") " pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.147067 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds8dg\" (UniqueName: \"kubernetes.io/projected/4cb069b9-d659-481c-bbb5-616c4af71d65-kube-api-access-ds8dg\") pod \"neutron-db-create-5f2hw\" (UID: \"4cb069b9-d659-481c-bbb5-616c4af71d65\") " pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.147712 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdknx\" (UniqueName: \"kubernetes.io/projected/7a18e489-bc4f-462a-87da-280408bb99e2-kube-api-access-pdknx\") pod \"keystone-db-sync-4sm7k\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.210271 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.282503 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:46 crc kubenswrapper[4946]: I1203 07:11:46.308946 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:47 crc kubenswrapper[4946]: I1203 07:11:47.099896 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:11:47 crc kubenswrapper[4946]: I1203 07:11:47.171542 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784d65c867-79wht"] Dec 03 07:11:47 crc kubenswrapper[4946]: I1203 07:11:47.172256 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-784d65c867-79wht" podUID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerName="dnsmasq-dns" containerID="cri-o://739eb262b67e2f24bd3529b30f8a2ef7d1f0674089c1cd2b2752b1e3e063ec1e" gracePeriod=10 Dec 03 07:11:47 crc kubenswrapper[4946]: I1203 07:11:47.471571 4946 generic.go:334] "Generic (PLEG): container finished" podID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerID="739eb262b67e2f24bd3529b30f8a2ef7d1f0674089c1cd2b2752b1e3e063ec1e" exitCode=0 Dec 03 07:11:47 crc kubenswrapper[4946]: I1203 07:11:47.471622 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784d65c867-79wht" event={"ID":"48843445-ca95-4a91-bf4b-05c16ebc945c","Type":"ContainerDied","Data":"739eb262b67e2f24bd3529b30f8a2ef7d1f0674089c1cd2b2752b1e3e063ec1e"} Dec 03 07:11:50 crc kubenswrapper[4946]: I1203 07:11:50.869289 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-784d65c867-79wht" podUID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.040441 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.040982 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.106024 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.140118 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-sb\") pod \"48843445-ca95-4a91-bf4b-05c16ebc945c\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.140172 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-dns-svc\") pod \"48843445-ca95-4a91-bf4b-05c16ebc945c\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.140259 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4d27\" (UniqueName: \"kubernetes.io/projected/48843445-ca95-4a91-bf4b-05c16ebc945c-kube-api-access-f4d27\") pod \"48843445-ca95-4a91-bf4b-05c16ebc945c\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.140277 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-nb\") pod \"48843445-ca95-4a91-bf4b-05c16ebc945c\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.140305 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-config\") pod \"48843445-ca95-4a91-bf4b-05c16ebc945c\" (UID: \"48843445-ca95-4a91-bf4b-05c16ebc945c\") " Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.144825 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48843445-ca95-4a91-bf4b-05c16ebc945c-kube-api-access-f4d27" (OuterVolumeSpecName: "kube-api-access-f4d27") pod "48843445-ca95-4a91-bf4b-05c16ebc945c" (UID: "48843445-ca95-4a91-bf4b-05c16ebc945c"). InnerVolumeSpecName "kube-api-access-f4d27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.180913 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "48843445-ca95-4a91-bf4b-05c16ebc945c" (UID: "48843445-ca95-4a91-bf4b-05c16ebc945c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.181419 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "48843445-ca95-4a91-bf4b-05c16ebc945c" (UID: "48843445-ca95-4a91-bf4b-05c16ebc945c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.181995 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48843445-ca95-4a91-bf4b-05c16ebc945c" (UID: "48843445-ca95-4a91-bf4b-05c16ebc945c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.186167 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-config" (OuterVolumeSpecName: "config") pod "48843445-ca95-4a91-bf4b-05c16ebc945c" (UID: "48843445-ca95-4a91-bf4b-05c16ebc945c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.242361 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4d27\" (UniqueName: \"kubernetes.io/projected/48843445-ca95-4a91-bf4b-05c16ebc945c-kube-api-access-f4d27\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.242399 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.242411 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.242422 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.242434 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48843445-ca95-4a91-bf4b-05c16ebc945c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.325023 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5f2hw"] Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.470337 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6e16-account-create-update-c4t4n"] Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.479299 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xj4g5"] Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.495054 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-eae6-account-create-update-xrhjp"] Dec 03 07:11:53 crc kubenswrapper[4946]: W1203 07:11:53.498142 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ebfd4c9_aa01_4ab5_82d7_506d0b97b573.slice/crio-acdd2ca8bef7cb91922488f09f24c5a7293d1bda242eadf5e7ef759aa92f221e WatchSource:0}: Error finding container acdd2ca8bef7cb91922488f09f24c5a7293d1bda242eadf5e7ef759aa92f221e: Status 404 returned error can't find the container with id acdd2ca8bef7cb91922488f09f24c5a7293d1bda242eadf5e7ef759aa92f221e Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.501690 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-fftdj"] Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.511122 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4sm7k"] Dec 03 07:11:53 crc kubenswrapper[4946]: W1203 07:11:53.516616 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a18e489_bc4f_462a_87da_280408bb99e2.slice/crio-da8464b6bd5a90e36469848e6318e8455287b1d8a9037305b82614a1e3eb9e99 WatchSource:0}: Error finding container da8464b6bd5a90e36469848e6318e8455287b1d8a9037305b82614a1e3eb9e99: Status 404 returned error can't find the container with id da8464b6bd5a90e36469848e6318e8455287b1d8a9037305b82614a1e3eb9e99 Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.556948 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784d65c867-79wht" event={"ID":"48843445-ca95-4a91-bf4b-05c16ebc945c","Type":"ContainerDied","Data":"6fb8b1393fdd7fed0333560564387f168d4b885a114eb568fb4c475a4ea1bec6"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.557404 4946 scope.go:117] "RemoveContainer" containerID="739eb262b67e2f24bd3529b30f8a2ef7d1f0674089c1cd2b2752b1e3e063ec1e" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.556988 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784d65c867-79wht" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.561257 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5f2hw" event={"ID":"4cb069b9-d659-481c-bbb5-616c4af71d65","Type":"ContainerStarted","Data":"243d3ab281717f39767cc58f3cd2061d2e057089fdd1678c26dadc9e51f4de2d"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.561298 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5f2hw" event={"ID":"4cb069b9-d659-481c-bbb5-616c4af71d65","Type":"ContainerStarted","Data":"483b932d072229dbddccbe346032428f4e41f6451a45d0105719d247fe26f05d"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.562932 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eae6-account-create-update-xrhjp" event={"ID":"880093ae-378a-4b62-902b-35b216d6d022","Type":"ContainerStarted","Data":"f364d217402252d142ff4be3e3b8f6f8747a597ba9d5300136ac667244e5f9b4"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.567752 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4sm7k" event={"ID":"7a18e489-bc4f-462a-87da-280408bb99e2","Type":"ContainerStarted","Data":"da8464b6bd5a90e36469848e6318e8455287b1d8a9037305b82614a1e3eb9e99"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.573343 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fftdj" event={"ID":"cdc5e484-c93b-47bf-8144-e76a5db68bfd","Type":"ContainerStarted","Data":"f6180b402ef09f43c3638dab7b5a31970f0479fe30383dfc5328c1112ac680a6"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.575933 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9q4dr" event={"ID":"a59904e6-2de4-4bc6-b0b5-6f578470b7d6","Type":"ContainerStarted","Data":"87aa8a68b0871fd821afa7b8d5609c2fd2858545673a2ba5c62eede8cb96b325"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.577686 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xj4g5" event={"ID":"3b4bacb1-58ed-42a2-a98f-b871b1f1047f","Type":"ContainerStarted","Data":"b9703d2d15c3857eb69f51c1ccbe93b52f0efdbd7d942a18e240f6fc62e073eb"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.582817 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6e16-account-create-update-c4t4n" event={"ID":"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573","Type":"ContainerStarted","Data":"acdd2ca8bef7cb91922488f09f24c5a7293d1bda242eadf5e7ef759aa92f221e"} Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.596279 4946 scope.go:117] "RemoveContainer" containerID="0a71a58692dde04b294a34938de02f2640847f1901fd5f7af9a59255f404836e" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.599148 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-5f2hw" podStartSLOduration=8.599133867 podStartE2EDuration="8.599133867s" podCreationTimestamp="2025-12-03 07:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:11:53.58236558 +0000 UTC m=+1306.379055699" watchObservedRunningTime="2025-12-03 07:11:53.599133867 +0000 UTC m=+1306.395823996" Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.612884 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784d65c867-79wht"] Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.631075 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784d65c867-79wht"] Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.639224 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5125-account-create-update-c4zxx"] Dec 03 07:11:53 crc kubenswrapper[4946]: I1203 07:11:53.655271 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-9q4dr" podStartSLOduration=2.309841944 podStartE2EDuration="15.655255797s" podCreationTimestamp="2025-12-03 07:11:38 +0000 UTC" firstStartedPulling="2025-12-03 07:11:39.528698557 +0000 UTC m=+1292.325388686" lastFinishedPulling="2025-12-03 07:11:52.87411241 +0000 UTC m=+1305.670802539" observedRunningTime="2025-12-03 07:11:53.625496582 +0000 UTC m=+1306.422186691" watchObservedRunningTime="2025-12-03 07:11:53.655255797 +0000 UTC m=+1306.451945896" Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.607500 4946 generic.go:334] "Generic (PLEG): container finished" podID="5ebfd4c9-aa01-4ab5-82d7-506d0b97b573" containerID="1108f67de44b6b63d6dab499b1df82ca4a27a61bd91177dc82437eaf17006c2a" exitCode=0 Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.607674 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6e16-account-create-update-c4t4n" event={"ID":"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573","Type":"ContainerDied","Data":"1108f67de44b6b63d6dab499b1df82ca4a27a61bd91177dc82437eaf17006c2a"} Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.613401 4946 generic.go:334] "Generic (PLEG): container finished" podID="4cb069b9-d659-481c-bbb5-616c4af71d65" containerID="243d3ab281717f39767cc58f3cd2061d2e057089fdd1678c26dadc9e51f4de2d" exitCode=0 Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.613463 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5f2hw" event={"ID":"4cb069b9-d659-481c-bbb5-616c4af71d65","Type":"ContainerDied","Data":"243d3ab281717f39767cc58f3cd2061d2e057089fdd1678c26dadc9e51f4de2d"} Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.615228 4946 generic.go:334] "Generic (PLEG): container finished" podID="880093ae-378a-4b62-902b-35b216d6d022" containerID="4ce9f26033ad21dcb676e42f79ece97ce0b09051bf144ebdb9aea2ae4003da81" exitCode=0 Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.615284 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eae6-account-create-update-xrhjp" event={"ID":"880093ae-378a-4b62-902b-35b216d6d022","Type":"ContainerDied","Data":"4ce9f26033ad21dcb676e42f79ece97ce0b09051bf144ebdb9aea2ae4003da81"} Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.618477 4946 generic.go:334] "Generic (PLEG): container finished" podID="cdc5e484-c93b-47bf-8144-e76a5db68bfd" containerID="a4ff77b363d0d7b530a95ecccec7f1113feeb3d0a4ee6fc234f8f36f96f15956" exitCode=0 Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.618562 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fftdj" event={"ID":"cdc5e484-c93b-47bf-8144-e76a5db68bfd","Type":"ContainerDied","Data":"a4ff77b363d0d7b530a95ecccec7f1113feeb3d0a4ee6fc234f8f36f96f15956"} Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.620715 4946 generic.go:334] "Generic (PLEG): container finished" podID="e282365b-8ff7-4fd8-9220-ccdac9865543" containerID="d20e227fd20281f8afac7a5f2b06e1522069ec70539b7c99d3e1814e068f79e7" exitCode=0 Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.620821 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5125-account-create-update-c4zxx" event={"ID":"e282365b-8ff7-4fd8-9220-ccdac9865543","Type":"ContainerDied","Data":"d20e227fd20281f8afac7a5f2b06e1522069ec70539b7c99d3e1814e068f79e7"} Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.620840 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5125-account-create-update-c4zxx" event={"ID":"e282365b-8ff7-4fd8-9220-ccdac9865543","Type":"ContainerStarted","Data":"fb0e70b0f116ea72384dc02e008f5c0cde04e47ba057f6e46f00067c336084a0"} Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.625218 4946 generic.go:334] "Generic (PLEG): container finished" podID="3b4bacb1-58ed-42a2-a98f-b871b1f1047f" containerID="740fb16d62aafc273f14ed8ada7248138673e6e2b5d0bfcb331e6f2fed1a3dde" exitCode=0 Dec 03 07:11:54 crc kubenswrapper[4946]: I1203 07:11:54.625417 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xj4g5" event={"ID":"3b4bacb1-58ed-42a2-a98f-b871b1f1047f","Type":"ContainerDied","Data":"740fb16d62aafc273f14ed8ada7248138673e6e2b5d0bfcb331e6f2fed1a3dde"} Dec 03 07:11:55 crc kubenswrapper[4946]: I1203 07:11:55.603259 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48843445-ca95-4a91-bf4b-05c16ebc945c" path="/var/lib/kubelet/pods/48843445-ca95-4a91-bf4b-05c16ebc945c/volumes" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.027299 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.056830 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.083236 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.088354 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.115219 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.119021 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.174283 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e282365b-8ff7-4fd8-9220-ccdac9865543-operator-scripts\") pod \"e282365b-8ff7-4fd8-9220-ccdac9865543\" (UID: \"e282365b-8ff7-4fd8-9220-ccdac9865543\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.174399 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880093ae-378a-4b62-902b-35b216d6d022-operator-scripts\") pod \"880093ae-378a-4b62-902b-35b216d6d022\" (UID: \"880093ae-378a-4b62-902b-35b216d6d022\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.174490 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slrls\" (UniqueName: \"kubernetes.io/projected/880093ae-378a-4b62-902b-35b216d6d022-kube-api-access-slrls\") pod \"880093ae-378a-4b62-902b-35b216d6d022\" (UID: \"880093ae-378a-4b62-902b-35b216d6d022\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.174551 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lk82\" (UniqueName: \"kubernetes.io/projected/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-kube-api-access-8lk82\") pod \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\" (UID: \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.174598 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-operator-scripts\") pod \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\" (UID: \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.174621 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwxp8\" (UniqueName: \"kubernetes.io/projected/e282365b-8ff7-4fd8-9220-ccdac9865543-kube-api-access-wwxp8\") pod \"e282365b-8ff7-4fd8-9220-ccdac9865543\" (UID: \"e282365b-8ff7-4fd8-9220-ccdac9865543\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.174675 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpx8v\" (UniqueName: \"kubernetes.io/projected/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-kube-api-access-xpx8v\") pod \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\" (UID: \"3b4bacb1-58ed-42a2-a98f-b871b1f1047f\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.174701 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-operator-scripts\") pod \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\" (UID: \"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.175171 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e282365b-8ff7-4fd8-9220-ccdac9865543-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e282365b-8ff7-4fd8-9220-ccdac9865543" (UID: "e282365b-8ff7-4fd8-9220-ccdac9865543"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.175577 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/880093ae-378a-4b62-902b-35b216d6d022-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "880093ae-378a-4b62-902b-35b216d6d022" (UID: "880093ae-378a-4b62-902b-35b216d6d022"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.175671 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ebfd4c9-aa01-4ab5-82d7-506d0b97b573" (UID: "5ebfd4c9-aa01-4ab5-82d7-506d0b97b573"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.176102 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3b4bacb1-58ed-42a2-a98f-b871b1f1047f" (UID: "3b4bacb1-58ed-42a2-a98f-b871b1f1047f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.181137 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e282365b-8ff7-4fd8-9220-ccdac9865543-kube-api-access-wwxp8" (OuterVolumeSpecName: "kube-api-access-wwxp8") pod "e282365b-8ff7-4fd8-9220-ccdac9865543" (UID: "e282365b-8ff7-4fd8-9220-ccdac9865543"). InnerVolumeSpecName "kube-api-access-wwxp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.182878 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880093ae-378a-4b62-902b-35b216d6d022-kube-api-access-slrls" (OuterVolumeSpecName: "kube-api-access-slrls") pod "880093ae-378a-4b62-902b-35b216d6d022" (UID: "880093ae-378a-4b62-902b-35b216d6d022"). InnerVolumeSpecName "kube-api-access-slrls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.183484 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-kube-api-access-xpx8v" (OuterVolumeSpecName: "kube-api-access-xpx8v") pod "3b4bacb1-58ed-42a2-a98f-b871b1f1047f" (UID: "3b4bacb1-58ed-42a2-a98f-b871b1f1047f"). InnerVolumeSpecName "kube-api-access-xpx8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.202056 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-kube-api-access-8lk82" (OuterVolumeSpecName: "kube-api-access-8lk82") pod "5ebfd4c9-aa01-4ab5-82d7-506d0b97b573" (UID: "5ebfd4c9-aa01-4ab5-82d7-506d0b97b573"). InnerVolumeSpecName "kube-api-access-8lk82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.276785 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99tjh\" (UniqueName: \"kubernetes.io/projected/cdc5e484-c93b-47bf-8144-e76a5db68bfd-kube-api-access-99tjh\") pod \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\" (UID: \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.276940 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb069b9-d659-481c-bbb5-616c4af71d65-operator-scripts\") pod \"4cb069b9-d659-481c-bbb5-616c4af71d65\" (UID: \"4cb069b9-d659-481c-bbb5-616c4af71d65\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.277579 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cb069b9-d659-481c-bbb5-616c4af71d65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4cb069b9-d659-481c-bbb5-616c4af71d65" (UID: "4cb069b9-d659-481c-bbb5-616c4af71d65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.277713 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds8dg\" (UniqueName: \"kubernetes.io/projected/4cb069b9-d659-481c-bbb5-616c4af71d65-kube-api-access-ds8dg\") pod \"4cb069b9-d659-481c-bbb5-616c4af71d65\" (UID: \"4cb069b9-d659-481c-bbb5-616c4af71d65\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.278294 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc5e484-c93b-47bf-8144-e76a5db68bfd-operator-scripts\") pod \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\" (UID: \"cdc5e484-c93b-47bf-8144-e76a5db68bfd\") " Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.278972 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdc5e484-c93b-47bf-8144-e76a5db68bfd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cdc5e484-c93b-47bf-8144-e76a5db68bfd" (UID: "cdc5e484-c93b-47bf-8144-e76a5db68bfd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279418 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc5e484-c93b-47bf-8144-e76a5db68bfd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279445 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/880093ae-378a-4b62-902b-35b216d6d022-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279459 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slrls\" (UniqueName: \"kubernetes.io/projected/880093ae-378a-4b62-902b-35b216d6d022-kube-api-access-slrls\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279473 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lk82\" (UniqueName: \"kubernetes.io/projected/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-kube-api-access-8lk82\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279484 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb069b9-d659-481c-bbb5-616c4af71d65-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279495 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279505 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwxp8\" (UniqueName: \"kubernetes.io/projected/e282365b-8ff7-4fd8-9220-ccdac9865543-kube-api-access-wwxp8\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279517 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpx8v\" (UniqueName: \"kubernetes.io/projected/3b4bacb1-58ed-42a2-a98f-b871b1f1047f-kube-api-access-xpx8v\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279528 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.279539 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e282365b-8ff7-4fd8-9220-ccdac9865543-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.280301 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdc5e484-c93b-47bf-8144-e76a5db68bfd-kube-api-access-99tjh" (OuterVolumeSpecName: "kube-api-access-99tjh") pod "cdc5e484-c93b-47bf-8144-e76a5db68bfd" (UID: "cdc5e484-c93b-47bf-8144-e76a5db68bfd"). InnerVolumeSpecName "kube-api-access-99tjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.280730 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cb069b9-d659-481c-bbb5-616c4af71d65-kube-api-access-ds8dg" (OuterVolumeSpecName: "kube-api-access-ds8dg") pod "4cb069b9-d659-481c-bbb5-616c4af71d65" (UID: "4cb069b9-d659-481c-bbb5-616c4af71d65"). InnerVolumeSpecName "kube-api-access-ds8dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.381802 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99tjh\" (UniqueName: \"kubernetes.io/projected/cdc5e484-c93b-47bf-8144-e76a5db68bfd-kube-api-access-99tjh\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.381846 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds8dg\" (UniqueName: \"kubernetes.io/projected/4cb069b9-d659-481c-bbb5-616c4af71d65-kube-api-access-ds8dg\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.667939 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5f2hw" event={"ID":"4cb069b9-d659-481c-bbb5-616c4af71d65","Type":"ContainerDied","Data":"483b932d072229dbddccbe346032428f4e41f6451a45d0105719d247fe26f05d"} Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.667999 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="483b932d072229dbddccbe346032428f4e41f6451a45d0105719d247fe26f05d" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.668054 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5f2hw" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.671159 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eae6-account-create-update-xrhjp" event={"ID":"880093ae-378a-4b62-902b-35b216d6d022","Type":"ContainerDied","Data":"f364d217402252d142ff4be3e3b8f6f8747a597ba9d5300136ac667244e5f9b4"} Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.671206 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f364d217402252d142ff4be3e3b8f6f8747a597ba9d5300136ac667244e5f9b4" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.671263 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eae6-account-create-update-xrhjp" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.673914 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4sm7k" event={"ID":"7a18e489-bc4f-462a-87da-280408bb99e2","Type":"ContainerStarted","Data":"6f7b1944bbb54c5a21449cb0e33457d34fcb8a0ab8e6a2a481e6fb4b73ed96cf"} Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.676984 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fftdj" event={"ID":"cdc5e484-c93b-47bf-8144-e76a5db68bfd","Type":"ContainerDied","Data":"f6180b402ef09f43c3638dab7b5a31970f0479fe30383dfc5328c1112ac680a6"} Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.677037 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fftdj" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.677066 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6180b402ef09f43c3638dab7b5a31970f0479fe30383dfc5328c1112ac680a6" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.679544 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5125-account-create-update-c4zxx" event={"ID":"e282365b-8ff7-4fd8-9220-ccdac9865543","Type":"ContainerDied","Data":"fb0e70b0f116ea72384dc02e008f5c0cde04e47ba057f6e46f00067c336084a0"} Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.679566 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb0e70b0f116ea72384dc02e008f5c0cde04e47ba057f6e46f00067c336084a0" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.679618 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5125-account-create-update-c4zxx" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.682889 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xj4g5" event={"ID":"3b4bacb1-58ed-42a2-a98f-b871b1f1047f","Type":"ContainerDied","Data":"b9703d2d15c3857eb69f51c1ccbe93b52f0efdbd7d942a18e240f6fc62e073eb"} Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.682914 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9703d2d15c3857eb69f51c1ccbe93b52f0efdbd7d942a18e240f6fc62e073eb" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.682888 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xj4g5" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.685839 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6e16-account-create-update-c4t4n" event={"ID":"5ebfd4c9-aa01-4ab5-82d7-506d0b97b573","Type":"ContainerDied","Data":"acdd2ca8bef7cb91922488f09f24c5a7293d1bda242eadf5e7ef759aa92f221e"} Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.685917 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acdd2ca8bef7cb91922488f09f24c5a7293d1bda242eadf5e7ef759aa92f221e" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.685982 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6e16-account-create-update-c4t4n" Dec 03 07:11:58 crc kubenswrapper[4946]: I1203 07:11:58.703727 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-4sm7k" podStartSLOduration=9.314017913 podStartE2EDuration="13.703703121s" podCreationTimestamp="2025-12-03 07:11:45 +0000 UTC" firstStartedPulling="2025-12-03 07:11:53.53298 +0000 UTC m=+1306.329670109" lastFinishedPulling="2025-12-03 07:11:57.922665198 +0000 UTC m=+1310.719355317" observedRunningTime="2025-12-03 07:11:58.701479711 +0000 UTC m=+1311.498169860" watchObservedRunningTime="2025-12-03 07:11:58.703703121 +0000 UTC m=+1311.500393270" Dec 03 07:12:01 crc kubenswrapper[4946]: I1203 07:12:01.714757 4946 generic.go:334] "Generic (PLEG): container finished" podID="7a18e489-bc4f-462a-87da-280408bb99e2" containerID="6f7b1944bbb54c5a21449cb0e33457d34fcb8a0ab8e6a2a481e6fb4b73ed96cf" exitCode=0 Dec 03 07:12:01 crc kubenswrapper[4946]: I1203 07:12:01.714775 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4sm7k" event={"ID":"7a18e489-bc4f-462a-87da-280408bb99e2","Type":"ContainerDied","Data":"6f7b1944bbb54c5a21449cb0e33457d34fcb8a0ab8e6a2a481e6fb4b73ed96cf"} Dec 03 07:12:01 crc kubenswrapper[4946]: I1203 07:12:01.717280 4946 generic.go:334] "Generic (PLEG): container finished" podID="a59904e6-2de4-4bc6-b0b5-6f578470b7d6" containerID="87aa8a68b0871fd821afa7b8d5609c2fd2858545673a2ba5c62eede8cb96b325" exitCode=0 Dec 03 07:12:01 crc kubenswrapper[4946]: I1203 07:12:01.717317 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9q4dr" event={"ID":"a59904e6-2de4-4bc6-b0b5-6f578470b7d6","Type":"ContainerDied","Data":"87aa8a68b0871fd821afa7b8d5609c2fd2858545673a2ba5c62eede8cb96b325"} Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.163714 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.169662 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9q4dr" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.268161 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-config-data\") pod \"7a18e489-bc4f-462a-87da-280408bb99e2\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.268330 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-combined-ca-bundle\") pod \"7a18e489-bc4f-462a-87da-280408bb99e2\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.268406 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdknx\" (UniqueName: \"kubernetes.io/projected/7a18e489-bc4f-462a-87da-280408bb99e2-kube-api-access-pdknx\") pod \"7a18e489-bc4f-462a-87da-280408bb99e2\" (UID: \"7a18e489-bc4f-462a-87da-280408bb99e2\") " Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.268558 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg4pt\" (UniqueName: \"kubernetes.io/projected/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-kube-api-access-pg4pt\") pod \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.274947 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-kube-api-access-pg4pt" (OuterVolumeSpecName: "kube-api-access-pg4pt") pod "a59904e6-2de4-4bc6-b0b5-6f578470b7d6" (UID: "a59904e6-2de4-4bc6-b0b5-6f578470b7d6"). InnerVolumeSpecName "kube-api-access-pg4pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.275252 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a18e489-bc4f-462a-87da-280408bb99e2-kube-api-access-pdknx" (OuterVolumeSpecName: "kube-api-access-pdknx") pod "7a18e489-bc4f-462a-87da-280408bb99e2" (UID: "7a18e489-bc4f-462a-87da-280408bb99e2"). InnerVolumeSpecName "kube-api-access-pdknx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.298496 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a18e489-bc4f-462a-87da-280408bb99e2" (UID: "7a18e489-bc4f-462a-87da-280408bb99e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.322456 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-config-data" (OuterVolumeSpecName: "config-data") pod "7a18e489-bc4f-462a-87da-280408bb99e2" (UID: "7a18e489-bc4f-462a-87da-280408bb99e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.370059 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-combined-ca-bundle\") pod \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.370197 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-config-data\") pod \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.370347 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-db-sync-config-data\") pod \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\" (UID: \"a59904e6-2de4-4bc6-b0b5-6f578470b7d6\") " Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.370843 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.370871 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdknx\" (UniqueName: \"kubernetes.io/projected/7a18e489-bc4f-462a-87da-280408bb99e2-kube-api-access-pdknx\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.370887 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg4pt\" (UniqueName: \"kubernetes.io/projected/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-kube-api-access-pg4pt\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.370901 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a18e489-bc4f-462a-87da-280408bb99e2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.375419 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a59904e6-2de4-4bc6-b0b5-6f578470b7d6" (UID: "a59904e6-2de4-4bc6-b0b5-6f578470b7d6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.391783 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a59904e6-2de4-4bc6-b0b5-6f578470b7d6" (UID: "a59904e6-2de4-4bc6-b0b5-6f578470b7d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.407007 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-config-data" (OuterVolumeSpecName: "config-data") pod "a59904e6-2de4-4bc6-b0b5-6f578470b7d6" (UID: "a59904e6-2de4-4bc6-b0b5-6f578470b7d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.472606 4946 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.472641 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.472653 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a59904e6-2de4-4bc6-b0b5-6f578470b7d6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.741841 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9q4dr" event={"ID":"a59904e6-2de4-4bc6-b0b5-6f578470b7d6","Type":"ContainerDied","Data":"45f1af885a6ef72f093d1d6772635efcacf75ddbaf5edba61bbd007f700eba72"} Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.741879 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45f1af885a6ef72f093d1d6772635efcacf75ddbaf5edba61bbd007f700eba72" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.741940 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9q4dr" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.744690 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4sm7k" event={"ID":"7a18e489-bc4f-462a-87da-280408bb99e2","Type":"ContainerDied","Data":"da8464b6bd5a90e36469848e6318e8455287b1d8a9037305b82614a1e3eb9e99"} Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.744713 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da8464b6bd5a90e36469848e6318e8455287b1d8a9037305b82614a1e3eb9e99" Dec 03 07:12:03 crc kubenswrapper[4946]: I1203 07:12:03.744775 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4sm7k" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.037447 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5j9vt"] Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.037867 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a18e489-bc4f-462a-87da-280408bb99e2" containerName="keystone-db-sync" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.037887 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a18e489-bc4f-462a-87da-280408bb99e2" containerName="keystone-db-sync" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.037898 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ebfd4c9-aa01-4ab5-82d7-506d0b97b573" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.037907 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ebfd4c9-aa01-4ab5-82d7-506d0b97b573" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.037916 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e282365b-8ff7-4fd8-9220-ccdac9865543" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.037924 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e282365b-8ff7-4fd8-9220-ccdac9865543" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.037942 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc5e484-c93b-47bf-8144-e76a5db68bfd" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.037950 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc5e484-c93b-47bf-8144-e76a5db68bfd" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.037972 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b4bacb1-58ed-42a2-a98f-b871b1f1047f" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.037983 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b4bacb1-58ed-42a2-a98f-b871b1f1047f" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.037999 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a59904e6-2de4-4bc6-b0b5-6f578470b7d6" containerName="glance-db-sync" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038006 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a59904e6-2de4-4bc6-b0b5-6f578470b7d6" containerName="glance-db-sync" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.038018 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerName="dnsmasq-dns" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038025 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerName="dnsmasq-dns" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.038035 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb069b9-d659-481c-bbb5-616c4af71d65" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038042 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb069b9-d659-481c-bbb5-616c4af71d65" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.038055 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880093ae-378a-4b62-902b-35b216d6d022" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038062 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="880093ae-378a-4b62-902b-35b216d6d022" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.038076 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerName="init" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038083 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerName="init" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038271 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ebfd4c9-aa01-4ab5-82d7-506d0b97b573" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038284 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b4bacb1-58ed-42a2-a98f-b871b1f1047f" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038294 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a59904e6-2de4-4bc6-b0b5-6f578470b7d6" containerName="glance-db-sync" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038307 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a18e489-bc4f-462a-87da-280408bb99e2" containerName="keystone-db-sync" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038325 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="48843445-ca95-4a91-bf4b-05c16ebc945c" containerName="dnsmasq-dns" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038340 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cb069b9-d659-481c-bbb5-616c4af71d65" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038351 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="e282365b-8ff7-4fd8-9220-ccdac9865543" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038366 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="880093ae-378a-4b62-902b-35b216d6d022" containerName="mariadb-account-create-update" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.038380 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc5e484-c93b-47bf-8144-e76a5db68bfd" containerName="mariadb-database-create" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.039041 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.047943 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.048134 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.048837 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mwdjs" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.053898 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.054676 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.062912 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5j9vt"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.086391 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-scripts\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.086666 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-credential-keys\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.086687 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-combined-ca-bundle\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.086706 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8lrt\" (UniqueName: \"kubernetes.io/projected/04271b31-6e46-456c-8efe-5c0c00b55650-kube-api-access-h8lrt\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.086854 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-config-data\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.086907 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-fernet-keys\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.103388 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f44b464f-d7k84"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.104621 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.116353 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f44b464f-d7k84"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.188246 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-fernet-keys\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.188308 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-scripts\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.188336 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-credential-keys\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.188367 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-combined-ca-bundle\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.188393 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8lrt\" (UniqueName: \"kubernetes.io/projected/04271b31-6e46-456c-8efe-5c0c00b55650-kube-api-access-h8lrt\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.188497 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-config-data\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.192639 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-config-data\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.209209 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-fernet-keys\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.209488 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-scripts\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.210010 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-combined-ca-bundle\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.210319 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-credential-keys\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.229306 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8lrt\" (UniqueName: \"kubernetes.io/projected/04271b31-6e46-456c-8efe-5c0c00b55650-kube-api-access-h8lrt\") pod \"keystone-bootstrap-5j9vt\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.251824 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-9twd7"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.253071 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.262521 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.262547 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-c2hp5" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.263274 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.269448 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-qzv4q"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.271260 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.285843 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-9twd7"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.305085 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-q8hq6" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.307601 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.307996 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.317201 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-swift-storage-0\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.317270 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc44j\" (UniqueName: \"kubernetes.io/projected/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-kube-api-access-mc44j\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.317413 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-sb\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.317580 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-config\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.318708 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-nb\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.319007 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-svc\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.321131 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qzv4q"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.414805 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.445876 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-db-sync-config-data\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.445921 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-sb\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.445940 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x8g8\" (UniqueName: \"kubernetes.io/projected/8e368226-3cfe-42a1-b2a3-52a53d227807-kube-api-access-8x8g8\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.445966 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-config\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.445989 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e368226-3cfe-42a1-b2a3-52a53d227807-etc-machine-id\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446016 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-config\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446036 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-nb\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446051 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-svc\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446073 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-combined-ca-bundle\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446097 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-combined-ca-bundle\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446118 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-scripts\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446147 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggr95\" (UniqueName: \"kubernetes.io/projected/6cba79e5-c505-4c72-9609-48838c0f59c5-kube-api-access-ggr95\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446172 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-swift-storage-0\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446191 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc44j\" (UniqueName: \"kubernetes.io/projected/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-kube-api-access-mc44j\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.446210 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-config-data\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.447068 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-nb\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.447384 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-swift-storage-0\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.447501 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-config\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.447586 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-sb\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.449535 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-svc\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.469878 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.471876 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.475114 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc44j\" (UniqueName: \"kubernetes.io/projected/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-kube-api-access-mc44j\") pod \"dnsmasq-dns-f44b464f-d7k84\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.480699 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.480905 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.493957 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.506789 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f44b464f-d7k84"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.507325 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.526081 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-qnf4v"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.527069 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.530356 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.530481 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rkhgb" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.547530 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qnf4v"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550595 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-config-data\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550656 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-db-sync-config-data\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550685 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x8g8\" (UniqueName: \"kubernetes.io/projected/8e368226-3cfe-42a1-b2a3-52a53d227807-kube-api-access-8x8g8\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550713 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-config\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550756 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e368226-3cfe-42a1-b2a3-52a53d227807-etc-machine-id\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550807 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-combined-ca-bundle\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550833 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-combined-ca-bundle\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550859 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-scripts\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.550895 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggr95\" (UniqueName: \"kubernetes.io/projected/6cba79e5-c505-4c72-9609-48838c0f59c5-kube-api-access-ggr95\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.552039 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e368226-3cfe-42a1-b2a3-52a53d227807-etc-machine-id\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.559284 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-db-sync-config-data\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.582434 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-combined-ca-bundle\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.583445 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-config\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.583532 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-config-data\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.583820 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-scripts\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.586367 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-combined-ca-bundle\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.596712 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c4d44c4d5-sjbvh"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.598055 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.601951 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggr95\" (UniqueName: \"kubernetes.io/projected/6cba79e5-c505-4c72-9609-48838c0f59c5-kube-api-access-ggr95\") pod \"neutron-db-sync-qzv4q\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.608049 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x8g8\" (UniqueName: \"kubernetes.io/projected/8e368226-3cfe-42a1-b2a3-52a53d227807-kube-api-access-8x8g8\") pod \"cinder-db-sync-9twd7\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.627636 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c4d44c4d5-sjbvh"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.653643 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-config-data\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.653707 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-combined-ca-bundle\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.653755 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm8pc\" (UniqueName: \"kubernetes.io/projected/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-kube-api-access-xm8pc\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.653800 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-scripts\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.653921 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.654040 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x44l2\" (UniqueName: \"kubernetes.io/projected/db4ad7e1-ae04-4f3a-8594-2897d4052e32-kube-api-access-x44l2\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.654064 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.654082 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-run-httpd\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.654103 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-log-httpd\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.654134 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.654160 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-db-sync-config-data\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.654528 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qhs9v"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.655661 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.659462 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.659608 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xlgln" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.659783 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.697024 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c4d44c4d5-sjbvh"] Dec 03 07:12:04 crc kubenswrapper[4946]: E1203 07:12:04.697647 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-28bvz ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" podUID="a1a7fab1-5829-48e0-8606-bb873e686fb7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.712016 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qhs9v"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.723809 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bd5694b85-wl6d8"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.725208 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.744825 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd5694b85-wl6d8"] Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.755920 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-swift-storage-0\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.755971 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-config-data\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.755998 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-combined-ca-bundle\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756028 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm8pc\" (UniqueName: \"kubernetes.io/projected/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-kube-api-access-xm8pc\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756059 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-svc\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756080 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28bvz\" (UniqueName: \"kubernetes.io/projected/a1a7fab1-5829-48e0-8606-bb873e686fb7-kube-api-access-28bvz\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756108 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-scripts\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756130 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x44l2\" (UniqueName: \"kubernetes.io/projected/db4ad7e1-ae04-4f3a-8594-2897d4052e32-kube-api-access-x44l2\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756149 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-nb\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756172 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756188 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-run-httpd\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756208 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-log-httpd\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756236 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-sb\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756261 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756288 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-db-sync-config-data\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.756313 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-config\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.757965 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-run-httpd\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.758204 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-log-httpd\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.766428 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.775729 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-combined-ca-bundle\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.776854 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-config-data\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.779465 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-scripts\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.779631 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.779972 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.785469 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x44l2\" (UniqueName: \"kubernetes.io/projected/db4ad7e1-ae04-4f3a-8594-2897d4052e32-kube-api-access-x44l2\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.786824 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-db-sync-config-data\") pod \"barbican-db-sync-qnf4v\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.787942 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm8pc\" (UniqueName: \"kubernetes.io/projected/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-kube-api-access-xm8pc\") pod \"ceilometer-0\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.821668 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.838168 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.859947 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-sb\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860041 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-scripts\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860090 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-config\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860118 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-swift-storage-0\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860163 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860190 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-svc\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860209 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0190a925-4703-4ae4-8ab4-150aef815c52-logs\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860259 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-config\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860283 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-config-data\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860319 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-combined-ca-bundle\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860345 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-svc\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860366 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9t2d\" (UniqueName: \"kubernetes.io/projected/1c8deacd-463c-496a-8521-de378c09ba06-kube-api-access-d9t2d\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860474 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860552 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28bvz\" (UniqueName: \"kubernetes.io/projected/a1a7fab1-5829-48e0-8606-bb873e686fb7-kube-api-access-28bvz\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860593 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860637 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-nb\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.860669 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwmw2\" (UniqueName: \"kubernetes.io/projected/0190a925-4703-4ae4-8ab4-150aef815c52-kube-api-access-fwmw2\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.861550 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-config\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.861853 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-svc\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.861842 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-sb\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.862544 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-nb\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.862549 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-swift-storage-0\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.878717 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28bvz\" (UniqueName: \"kubernetes.io/projected/a1a7fab1-5829-48e0-8606-bb873e686fb7-kube-api-access-28bvz\") pod \"dnsmasq-dns-7c4d44c4d5-sjbvh\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.897193 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.908126 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961366 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-sb\") pod \"a1a7fab1-5829-48e0-8606-bb873e686fb7\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961448 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-svc\") pod \"a1a7fab1-5829-48e0-8606-bb873e686fb7\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961531 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-nb\") pod \"a1a7fab1-5829-48e0-8606-bb873e686fb7\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961586 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-config\") pod \"a1a7fab1-5829-48e0-8606-bb873e686fb7\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961609 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-swift-storage-0\") pod \"a1a7fab1-5829-48e0-8606-bb873e686fb7\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961862 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961904 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwmw2\" (UniqueName: \"kubernetes.io/projected/0190a925-4703-4ae4-8ab4-150aef815c52-kube-api-access-fwmw2\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961955 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-scripts\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.961986 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962007 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-svc\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962026 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0190a925-4703-4ae4-8ab4-150aef815c52-logs\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962049 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-config\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962072 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-config-data\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962087 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-combined-ca-bundle\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962092 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a1a7fab1-5829-48e0-8606-bb873e686fb7" (UID: "a1a7fab1-5829-48e0-8606-bb873e686fb7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962109 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9t2d\" (UniqueName: \"kubernetes.io/projected/1c8deacd-463c-496a-8521-de378c09ba06-kube-api-access-d9t2d\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962191 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962514 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962516 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a1a7fab1-5829-48e0-8606-bb873e686fb7" (UID: "a1a7fab1-5829-48e0-8606-bb873e686fb7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.962857 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a1a7fab1-5829-48e0-8606-bb873e686fb7" (UID: "a1a7fab1-5829-48e0-8606-bb873e686fb7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.963439 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.963453 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-config" (OuterVolumeSpecName: "config") pod "a1a7fab1-5829-48e0-8606-bb873e686fb7" (UID: "a1a7fab1-5829-48e0-8606-bb873e686fb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.964985 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-svc\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.965781 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a1a7fab1-5829-48e0-8606-bb873e686fb7" (UID: "a1a7fab1-5829-48e0-8606-bb873e686fb7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.966838 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.967699 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0190a925-4703-4ae4-8ab4-150aef815c52-logs\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.969015 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.971851 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-scripts\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.972128 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-config\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.977208 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-config-data\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.977773 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-combined-ca-bundle\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.980574 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwmw2\" (UniqueName: \"kubernetes.io/projected/0190a925-4703-4ae4-8ab4-150aef815c52-kube-api-access-fwmw2\") pod \"placement-db-sync-qhs9v\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.981979 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:04 crc kubenswrapper[4946]: I1203 07:12:04.984853 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9t2d\" (UniqueName: \"kubernetes.io/projected/1c8deacd-463c-496a-8521-de378c09ba06-kube-api-access-d9t2d\") pod \"dnsmasq-dns-7bd5694b85-wl6d8\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.017781 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5j9vt"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.068433 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28bvz\" (UniqueName: \"kubernetes.io/projected/a1a7fab1-5829-48e0-8606-bb873e686fb7-kube-api-access-28bvz\") pod \"a1a7fab1-5829-48e0-8606-bb873e686fb7\" (UID: \"a1a7fab1-5829-48e0-8606-bb873e686fb7\") " Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.068962 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.068974 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.068985 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.068994 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1a7fab1-5829-48e0-8606-bb873e686fb7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.071967 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a7fab1-5829-48e0-8606-bb873e686fb7-kube-api-access-28bvz" (OuterVolumeSpecName: "kube-api-access-28bvz") pod "a1a7fab1-5829-48e0-8606-bb873e686fb7" (UID: "a1a7fab1-5829-48e0-8606-bb873e686fb7"). InnerVolumeSpecName "kube-api-access-28bvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.126629 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.169934 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f44b464f-d7k84"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.170329 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28bvz\" (UniqueName: \"kubernetes.io/projected/a1a7fab1-5829-48e0-8606-bb873e686fb7-kube-api-access-28bvz\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:05 crc kubenswrapper[4946]: W1203 07:12:05.219210 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8e3aab2_4895_4e40_a3e6_4e4c18d58561.slice/crio-0a3b086f80d85f485212bacfbafd16fc872d4096de7605ae0fc1a19d49954382 WatchSource:0}: Error finding container 0a3b086f80d85f485212bacfbafd16fc872d4096de7605ae0fc1a19d49954382: Status 404 returned error can't find the container with id 0a3b086f80d85f485212bacfbafd16fc872d4096de7605ae0fc1a19d49954382 Dec 03 07:12:05 crc kubenswrapper[4946]: W1203 07:12:05.297851 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cba79e5_c505_4c72_9609_48838c0f59c5.slice/crio-782ba6b6aebd5af6d1efaa97f6845d2c93e271022e62716c9c6b3e77cc1ec021 WatchSource:0}: Error finding container 782ba6b6aebd5af6d1efaa97f6845d2c93e271022e62716c9c6b3e77cc1ec021: Status 404 returned error can't find the container with id 782ba6b6aebd5af6d1efaa97f6845d2c93e271022e62716c9c6b3e77cc1ec021 Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.305714 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qzv4q"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.319151 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.320636 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.324513 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.324687 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.324835 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bk8hz" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.330279 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.395129 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.481073 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jktj\" (UniqueName: \"kubernetes.io/projected/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-kube-api-access-7jktj\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.481141 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.481210 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-logs\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.481244 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.481275 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.481328 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-scripts\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.481371 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-config-data\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.587544 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-scripts\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.587602 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-config-data\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.587678 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jktj\" (UniqueName: \"kubernetes.io/projected/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-kube-api-access-7jktj\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.587710 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.587748 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-logs\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.587770 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.587789 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.588575 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.591954 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-logs\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.592995 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.596126 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.602286 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-config-data\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.602729 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-scripts\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.617321 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jktj\" (UniqueName: \"kubernetes.io/projected/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-kube-api-access-7jktj\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.638800 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.652456 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-9twd7"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.669400 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qhs9v"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.669618 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.689637 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qnf4v"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.699714 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.701618 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.703578 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.710021 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.789594 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qnf4v" event={"ID":"db4ad7e1-ae04-4f3a-8594-2897d4052e32","Type":"ContainerStarted","Data":"90f69fd381331cc338f5a8564f1d853a69322ce9f61fc3cc834d37b13077a628"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.790684 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.790715 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcpgp\" (UniqueName: \"kubernetes.io/projected/7740830b-2468-417a-9f4b-6458026070b6-kube-api-access-gcpgp\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.790870 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.790903 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.790936 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-logs\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.791220 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.791273 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.793537 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qzv4q" event={"ID":"6cba79e5-c505-4c72-9609-48838c0f59c5","Type":"ContainerStarted","Data":"8dcad6ca973112b601de433a0add9cb4252be0e542acfd5c99157c13bbe11c5a"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.793563 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qzv4q" event={"ID":"6cba79e5-c505-4c72-9609-48838c0f59c5","Type":"ContainerStarted","Data":"782ba6b6aebd5af6d1efaa97f6845d2c93e271022e62716c9c6b3e77cc1ec021"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.812647 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5j9vt" event={"ID":"04271b31-6e46-456c-8efe-5c0c00b55650","Type":"ContainerStarted","Data":"709c3aea255550cce7f8b4bca9af001a3c96f7d77f9a4aeadadf2f13c22bb5b5"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.812690 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5j9vt" event={"ID":"04271b31-6e46-456c-8efe-5c0c00b55650","Type":"ContainerStarted","Data":"6276a0d8b86e319b3cb60d397fcfb83cc2669d758063a3e9ef875df3c91e4e13"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.822817 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd5694b85-wl6d8"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.822851 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e32c0a4-449e-4f0f-9ab4-d020d313a99b","Type":"ContainerStarted","Data":"d8d84d715a68dabefe658a5385ed7fdb8da81a3f83b33daa97de451ac1fc8803"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.825968 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-qzv4q" podStartSLOduration=1.825951961 podStartE2EDuration="1.825951961s" podCreationTimestamp="2025-12-03 07:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:05.818192324 +0000 UTC m=+1318.614882433" watchObservedRunningTime="2025-12-03 07:12:05.825951961 +0000 UTC m=+1318.622642070" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.831495 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-9twd7" event={"ID":"8e368226-3cfe-42a1-b2a3-52a53d227807","Type":"ContainerStarted","Data":"2e7baa520f177e39464d54a6ddd50ea9677b11f3b55b8bcc888800e2769db65e"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.836838 4946 generic.go:334] "Generic (PLEG): container finished" podID="a8e3aab2-4895-4e40-a3e6-4e4c18d58561" containerID="97004b6f4aaa03cf401d52abe6fb51abe39e3bdbf334c5af80c20be6cb999048" exitCode=0 Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.836924 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f44b464f-d7k84" event={"ID":"a8e3aab2-4895-4e40-a3e6-4e4c18d58561","Type":"ContainerDied","Data":"97004b6f4aaa03cf401d52abe6fb51abe39e3bdbf334c5af80c20be6cb999048"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.836952 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f44b464f-d7k84" event={"ID":"a8e3aab2-4895-4e40-a3e6-4e4c18d58561","Type":"ContainerStarted","Data":"0a3b086f80d85f485212bacfbafd16fc872d4096de7605ae0fc1a19d49954382"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.849732 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c4d44c4d5-sjbvh" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.850779 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qhs9v" event={"ID":"0190a925-4703-4ae4-8ab4-150aef815c52","Type":"ContainerStarted","Data":"c7736f10fd024ff8119cc401925d35c4d22d580e40c4e2149d932f9904fb136d"} Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.875148 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5j9vt" podStartSLOduration=2.875109114 podStartE2EDuration="2.875109114s" podCreationTimestamp="2025-12-03 07:12:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:05.84690111 +0000 UTC m=+1318.643591249" watchObservedRunningTime="2025-12-03 07:12:05.875109114 +0000 UTC m=+1318.671799233" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.902062 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.902596 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.902680 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-logs\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.902806 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.902853 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.902966 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.903044 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcpgp\" (UniqueName: \"kubernetes.io/projected/7740830b-2468-417a-9f4b-6458026070b6-kube-api-access-gcpgp\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.903577 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.905870 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-logs\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.906160 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.911309 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.928721 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.940278 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcpgp\" (UniqueName: \"kubernetes.io/projected/7740830b-2468-417a-9f4b-6458026070b6-kube-api-access-gcpgp\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.946533 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.952598 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c4d44c4d5-sjbvh"] Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.976384 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:05 crc kubenswrapper[4946]: I1203 07:12:05.983548 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c4d44c4d5-sjbvh"] Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.034187 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.284186 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.360936 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.517926 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-svc\") pod \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.518586 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-sb\") pod \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.519482 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-nb\") pod \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.519684 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-config\") pod \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.519798 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-swift-storage-0\") pod \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.519851 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc44j\" (UniqueName: \"kubernetes.io/projected/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-kube-api-access-mc44j\") pod \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\" (UID: \"a8e3aab2-4895-4e40-a3e6-4e4c18d58561\") " Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.525594 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-kube-api-access-mc44j" (OuterVolumeSpecName: "kube-api-access-mc44j") pod "a8e3aab2-4895-4e40-a3e6-4e4c18d58561" (UID: "a8e3aab2-4895-4e40-a3e6-4e4c18d58561"). InnerVolumeSpecName "kube-api-access-mc44j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:06 crc kubenswrapper[4946]: I1203 07:12:06.550949 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a8e3aab2-4895-4e40-a3e6-4e4c18d58561" (UID: "a8e3aab2-4895-4e40-a3e6-4e4c18d58561"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:06 crc kubenswrapper[4946]: E1203 07:12:06.580472 4946 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c8deacd_463c_496a_8521_de378c09ba06.slice/crio-conmon-f8e7f32fa0ed4a638857013ca3e0d38b41eaadf0d05fcfc3fb9ad097dd5e8b11.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c8deacd_463c_496a_8521_de378c09ba06.slice/crio-f8e7f32fa0ed4a638857013ca3e0d38b41eaadf0d05fcfc3fb9ad097dd5e8b11.scope\": RecentStats: unable to find data in memory cache]" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.585956 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a8e3aab2-4895-4e40-a3e6-4e4c18d58561" (UID: "a8e3aab2-4895-4e40-a3e6-4e4c18d58561"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.586714 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a8e3aab2-4895-4e40-a3e6-4e4c18d58561" (UID: "a8e3aab2-4895-4e40-a3e6-4e4c18d58561"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.589074 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-config" (OuterVolumeSpecName: "config") pod "a8e3aab2-4895-4e40-a3e6-4e4c18d58561" (UID: "a8e3aab2-4895-4e40-a3e6-4e4c18d58561"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.608236 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a8e3aab2-4895-4e40-a3e6-4e4c18d58561" (UID: "a8e3aab2-4895-4e40-a3e6-4e4c18d58561"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.625319 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc44j\" (UniqueName: \"kubernetes.io/projected/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-kube-api-access-mc44j\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.625346 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.625358 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.625371 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.625383 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.625392 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8e3aab2-4895-4e40-a3e6-4e4c18d58561-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.660760 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.880516 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7740830b-2468-417a-9f4b-6458026070b6","Type":"ContainerStarted","Data":"1deffa94aef04dab39663d0ea4d237cde3bc583fa0735ddffa29762c492fec13"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.885899 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4","Type":"ContainerStarted","Data":"672546271f7aa5ad0f2f30b86637210d544510ce738d528555429c1b2db73fdc"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.894178 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f44b464f-d7k84" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.894174 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f44b464f-d7k84" event={"ID":"a8e3aab2-4895-4e40-a3e6-4e4c18d58561","Type":"ContainerDied","Data":"0a3b086f80d85f485212bacfbafd16fc872d4096de7605ae0fc1a19d49954382"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.894281 4946 scope.go:117] "RemoveContainer" containerID="97004b6f4aaa03cf401d52abe6fb51abe39e3bdbf334c5af80c20be6cb999048" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.898470 4946 generic.go:334] "Generic (PLEG): container finished" podID="1c8deacd-463c-496a-8521-de378c09ba06" containerID="f8e7f32fa0ed4a638857013ca3e0d38b41eaadf0d05fcfc3fb9ad097dd5e8b11" exitCode=0 Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.898515 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" event={"ID":"1c8deacd-463c-496a-8521-de378c09ba06","Type":"ContainerDied","Data":"f8e7f32fa0ed4a638857013ca3e0d38b41eaadf0d05fcfc3fb9ad097dd5e8b11"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.898538 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" event={"ID":"1c8deacd-463c-496a-8521-de378c09ba06","Type":"ContainerStarted","Data":"41c4df89e8715ff7049b467c3fb5d382eb9f1386265a7fb68b4ddea1388f73da"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:06.982790 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f44b464f-d7k84"] Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.013324 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f44b464f-d7k84"] Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.238944 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.279899 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.462671 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.623655 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1a7fab1-5829-48e0-8606-bb873e686fb7" path="/var/lib/kubelet/pods/a1a7fab1-5829-48e0-8606-bb873e686fb7/volumes" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.624128 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8e3aab2-4895-4e40-a3e6-4e4c18d58561" path="/var/lib/kubelet/pods/a8e3aab2-4895-4e40-a3e6-4e4c18d58561/volumes" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.957215 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" event={"ID":"1c8deacd-463c-496a-8521-de378c09ba06","Type":"ContainerStarted","Data":"41f29978308af79c28f1ea282a318bc72fb22a3e70de9e072d4cf64063188c1e"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.957331 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.959578 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7740830b-2468-417a-9f4b-6458026070b6","Type":"ContainerStarted","Data":"b378856018a35c42b4a69b49659e833c49da4683efdeeb418c5ff78884cf1b5a"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.962108 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4","Type":"ContainerStarted","Data":"d1d9b316ace5b8774924494a238c210e238707a050346e296ee5a326a63523cc"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:07.984016 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" podStartSLOduration=3.983999077 podStartE2EDuration="3.983999077s" podCreationTimestamp="2025-12-03 07:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:07.974418831 +0000 UTC m=+1320.771108940" watchObservedRunningTime="2025-12-03 07:12:07.983999077 +0000 UTC m=+1320.780689176" Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:08.974097 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4","Type":"ContainerStarted","Data":"07151b90e6fa289b1d4731210cc4334e3117596569ac5f36a61da46e73b5e09e"} Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:08.974188 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerName="glance-log" containerID="cri-o://d1d9b316ace5b8774924494a238c210e238707a050346e296ee5a326a63523cc" gracePeriod=30 Dec 03 07:12:08 crc kubenswrapper[4946]: I1203 07:12:08.974253 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerName="glance-httpd" containerID="cri-o://07151b90e6fa289b1d4731210cc4334e3117596569ac5f36a61da46e73b5e09e" gracePeriod=30 Dec 03 07:12:09 crc kubenswrapper[4946]: I1203 07:12:09.001682 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.00162168 podStartE2EDuration="5.00162168s" podCreationTimestamp="2025-12-03 07:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:08.992090466 +0000 UTC m=+1321.788780575" watchObservedRunningTime="2025-12-03 07:12:09.00162168 +0000 UTC m=+1321.798311789" Dec 03 07:12:09 crc kubenswrapper[4946]: I1203 07:12:09.004454 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7740830b-2468-417a-9f4b-6458026070b6","Type":"ContainerStarted","Data":"4ca4cd59686260af266576a5da095516a30bacdc7a29e37f3588e79be9586818"} Dec 03 07:12:09 crc kubenswrapper[4946]: I1203 07:12:09.004485 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7740830b-2468-417a-9f4b-6458026070b6" containerName="glance-log" containerID="cri-o://b378856018a35c42b4a69b49659e833c49da4683efdeeb418c5ff78884cf1b5a" gracePeriod=30 Dec 03 07:12:09 crc kubenswrapper[4946]: I1203 07:12:09.004579 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7740830b-2468-417a-9f4b-6458026070b6" containerName="glance-httpd" containerID="cri-o://4ca4cd59686260af266576a5da095516a30bacdc7a29e37f3588e79be9586818" gracePeriod=30 Dec 03 07:12:09 crc kubenswrapper[4946]: I1203 07:12:09.031011 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.030989935 podStartE2EDuration="5.030989935s" podCreationTimestamp="2025-12-03 07:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:09.024460371 +0000 UTC m=+1321.821150500" watchObservedRunningTime="2025-12-03 07:12:09.030989935 +0000 UTC m=+1321.827680044" Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.014032 4946 generic.go:334] "Generic (PLEG): container finished" podID="04271b31-6e46-456c-8efe-5c0c00b55650" containerID="709c3aea255550cce7f8b4bca9af001a3c96f7d77f9a4aeadadf2f13c22bb5b5" exitCode=0 Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.014214 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5j9vt" event={"ID":"04271b31-6e46-456c-8efe-5c0c00b55650","Type":"ContainerDied","Data":"709c3aea255550cce7f8b4bca9af001a3c96f7d77f9a4aeadadf2f13c22bb5b5"} Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.018213 4946 generic.go:334] "Generic (PLEG): container finished" podID="7740830b-2468-417a-9f4b-6458026070b6" containerID="4ca4cd59686260af266576a5da095516a30bacdc7a29e37f3588e79be9586818" exitCode=0 Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.018249 4946 generic.go:334] "Generic (PLEG): container finished" podID="7740830b-2468-417a-9f4b-6458026070b6" containerID="b378856018a35c42b4a69b49659e833c49da4683efdeeb418c5ff78884cf1b5a" exitCode=143 Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.018310 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7740830b-2468-417a-9f4b-6458026070b6","Type":"ContainerDied","Data":"4ca4cd59686260af266576a5da095516a30bacdc7a29e37f3588e79be9586818"} Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.018333 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7740830b-2468-417a-9f4b-6458026070b6","Type":"ContainerDied","Data":"b378856018a35c42b4a69b49659e833c49da4683efdeeb418c5ff78884cf1b5a"} Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.025409 4946 generic.go:334] "Generic (PLEG): container finished" podID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerID="07151b90e6fa289b1d4731210cc4334e3117596569ac5f36a61da46e73b5e09e" exitCode=0 Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.025441 4946 generic.go:334] "Generic (PLEG): container finished" podID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerID="d1d9b316ace5b8774924494a238c210e238707a050346e296ee5a326a63523cc" exitCode=143 Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.025463 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4","Type":"ContainerDied","Data":"07151b90e6fa289b1d4731210cc4334e3117596569ac5f36a61da46e73b5e09e"} Dec 03 07:12:10 crc kubenswrapper[4946]: I1203 07:12:10.025491 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4","Type":"ContainerDied","Data":"d1d9b316ace5b8774924494a238c210e238707a050346e296ee5a326a63523cc"} Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.411614 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.512541 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-credential-keys\") pod \"04271b31-6e46-456c-8efe-5c0c00b55650\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.512623 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-fernet-keys\") pod \"04271b31-6e46-456c-8efe-5c0c00b55650\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.512654 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-scripts\") pod \"04271b31-6e46-456c-8efe-5c0c00b55650\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.512831 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8lrt\" (UniqueName: \"kubernetes.io/projected/04271b31-6e46-456c-8efe-5c0c00b55650-kube-api-access-h8lrt\") pod \"04271b31-6e46-456c-8efe-5c0c00b55650\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.512872 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-combined-ca-bundle\") pod \"04271b31-6e46-456c-8efe-5c0c00b55650\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.512905 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-config-data\") pod \"04271b31-6e46-456c-8efe-5c0c00b55650\" (UID: \"04271b31-6e46-456c-8efe-5c0c00b55650\") " Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.518842 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "04271b31-6e46-456c-8efe-5c0c00b55650" (UID: "04271b31-6e46-456c-8efe-5c0c00b55650"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.521033 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04271b31-6e46-456c-8efe-5c0c00b55650-kube-api-access-h8lrt" (OuterVolumeSpecName: "kube-api-access-h8lrt") pod "04271b31-6e46-456c-8efe-5c0c00b55650" (UID: "04271b31-6e46-456c-8efe-5c0c00b55650"). InnerVolumeSpecName "kube-api-access-h8lrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.521084 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "04271b31-6e46-456c-8efe-5c0c00b55650" (UID: "04271b31-6e46-456c-8efe-5c0c00b55650"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.520994 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-scripts" (OuterVolumeSpecName: "scripts") pod "04271b31-6e46-456c-8efe-5c0c00b55650" (UID: "04271b31-6e46-456c-8efe-5c0c00b55650"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.541402 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04271b31-6e46-456c-8efe-5c0c00b55650" (UID: "04271b31-6e46-456c-8efe-5c0c00b55650"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.549520 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-config-data" (OuterVolumeSpecName: "config-data") pod "04271b31-6e46-456c-8efe-5c0c00b55650" (UID: "04271b31-6e46-456c-8efe-5c0c00b55650"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.615226 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8lrt\" (UniqueName: \"kubernetes.io/projected/04271b31-6e46-456c-8efe-5c0c00b55650-kube-api-access-h8lrt\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.615259 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.615271 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.615280 4946 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.615289 4946 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:13 crc kubenswrapper[4946]: I1203 07:12:13.615297 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04271b31-6e46-456c-8efe-5c0c00b55650-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.073424 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5j9vt" event={"ID":"04271b31-6e46-456c-8efe-5c0c00b55650","Type":"ContainerDied","Data":"6276a0d8b86e319b3cb60d397fcfb83cc2669d758063a3e9ef875df3c91e4e13"} Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.073469 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6276a0d8b86e319b3cb60d397fcfb83cc2669d758063a3e9ef875df3c91e4e13" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.073541 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5j9vt" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.509419 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5j9vt"] Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.517266 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5j9vt"] Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.622104 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rwzxk"] Dec 03 07:12:14 crc kubenswrapper[4946]: E1203 07:12:14.622464 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e3aab2-4895-4e40-a3e6-4e4c18d58561" containerName="init" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.622480 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e3aab2-4895-4e40-a3e6-4e4c18d58561" containerName="init" Dec 03 07:12:14 crc kubenswrapper[4946]: E1203 07:12:14.622516 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04271b31-6e46-456c-8efe-5c0c00b55650" containerName="keystone-bootstrap" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.622524 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="04271b31-6e46-456c-8efe-5c0c00b55650" containerName="keystone-bootstrap" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.622672 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8e3aab2-4895-4e40-a3e6-4e4c18d58561" containerName="init" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.622694 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="04271b31-6e46-456c-8efe-5c0c00b55650" containerName="keystone-bootstrap" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.624270 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.626813 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.627283 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.627357 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mwdjs" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.627443 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.631202 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.640427 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-combined-ca-bundle\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.640641 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-credential-keys\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.640822 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s2g2\" (UniqueName: \"kubernetes.io/projected/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-kube-api-access-4s2g2\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.640970 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-scripts\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.641132 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-fernet-keys\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.641242 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-config-data\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.654731 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rwzxk"] Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.743076 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-config-data\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.743200 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-combined-ca-bundle\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.743246 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-credential-keys\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.743312 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s2g2\" (UniqueName: \"kubernetes.io/projected/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-kube-api-access-4s2g2\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.743372 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-scripts\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.743416 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-fernet-keys\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.753484 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-config-data\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.754475 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-fernet-keys\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.761232 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-scripts\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.762475 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-combined-ca-bundle\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.770683 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-credential-keys\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.816773 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s2g2\" (UniqueName: \"kubernetes.io/projected/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-kube-api-access-4s2g2\") pod \"keystone-bootstrap-rwzxk\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:14 crc kubenswrapper[4946]: I1203 07:12:14.940953 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.128875 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.201191 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5cc849d9-2r4ds"] Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.201640 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" podUID="af0dc160-2364-4866-b68d-37fe8990241d" containerName="dnsmasq-dns" containerID="cri-o://f7fac8d05109eff647ba6132cda7c2e640a242149c274d302f87f99b27b94641" gracePeriod=10 Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.622212 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04271b31-6e46-456c-8efe-5c0c00b55650" path="/var/lib/kubelet/pods/04271b31-6e46-456c-8efe-5c0c00b55650/volumes" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.676847 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.761071 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.761469 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-scripts\") pod \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.761547 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-combined-ca-bundle\") pod \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.761595 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jktj\" (UniqueName: \"kubernetes.io/projected/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-kube-api-access-7jktj\") pod \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.761627 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-config-data\") pod \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.761905 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-logs\") pod \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.762036 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-httpd-run\") pod \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\" (UID: \"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4\") " Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.762846 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" (UID: "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.763065 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-logs" (OuterVolumeSpecName: "logs") pod "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" (UID: "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.768726 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-kube-api-access-7jktj" (OuterVolumeSpecName: "kube-api-access-7jktj") pod "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" (UID: "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4"). InnerVolumeSpecName "kube-api-access-7jktj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.798044 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" (UID: "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.798044 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-scripts" (OuterVolumeSpecName: "scripts") pod "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" (UID: "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.802206 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" (UID: "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.823970 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-config-data" (OuterVolumeSpecName: "config-data") pod "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" (UID: "003b7cd6-fdb9-4d1f-8159-41e7ec78caa4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.863937 4946 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.864026 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.864038 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.864048 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.864059 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jktj\" (UniqueName: \"kubernetes.io/projected/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-kube-api-access-7jktj\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.869847 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.869864 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.883677 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 07:12:15 crc kubenswrapper[4946]: I1203 07:12:15.971057 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.099060 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"003b7cd6-fdb9-4d1f-8159-41e7ec78caa4","Type":"ContainerDied","Data":"672546271f7aa5ad0f2f30b86637210d544510ce738d528555429c1b2db73fdc"} Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.099128 4946 scope.go:117] "RemoveContainer" containerID="07151b90e6fa289b1d4731210cc4334e3117596569ac5f36a61da46e73b5e09e" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.099087 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.101329 4946 generic.go:334] "Generic (PLEG): container finished" podID="af0dc160-2364-4866-b68d-37fe8990241d" containerID="f7fac8d05109eff647ba6132cda7c2e640a242149c274d302f87f99b27b94641" exitCode=0 Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.101370 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" event={"ID":"af0dc160-2364-4866-b68d-37fe8990241d","Type":"ContainerDied","Data":"f7fac8d05109eff647ba6132cda7c2e640a242149c274d302f87f99b27b94641"} Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.145019 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.155453 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.172179 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:16 crc kubenswrapper[4946]: E1203 07:12:16.172575 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerName="glance-httpd" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.172592 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerName="glance-httpd" Dec 03 07:12:16 crc kubenswrapper[4946]: E1203 07:12:16.172611 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerName="glance-log" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.172617 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerName="glance-log" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.172777 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerName="glance-httpd" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.172806 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" containerName="glance-log" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.174425 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.176603 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.181733 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.181986 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.275913 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.275972 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.276006 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7zmw\" (UniqueName: \"kubernetes.io/projected/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-kube-api-access-b7zmw\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.276033 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-logs\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.276166 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-config-data\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.276247 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.276279 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.276319 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-scripts\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.378918 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.378973 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.379008 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-scripts\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.379075 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.379107 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.379158 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7zmw\" (UniqueName: \"kubernetes.io/projected/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-kube-api-access-b7zmw\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.379182 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-logs\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.379238 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-config-data\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.379528 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.379898 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-logs\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.380100 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.386429 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-scripts\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.386833 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.395463 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.398134 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-config-data\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.400164 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7zmw\" (UniqueName: \"kubernetes.io/projected/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-kube-api-access-b7zmw\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.419796 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " pod="openstack/glance-default-external-api-0" Dec 03 07:12:16 crc kubenswrapper[4946]: I1203 07:12:16.493562 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:12:17 crc kubenswrapper[4946]: I1203 07:12:17.606651 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="003b7cd6-fdb9-4d1f-8159-41e7ec78caa4" path="/var/lib/kubelet/pods/003b7cd6-fdb9-4d1f-8159-41e7ec78caa4/volumes" Dec 03 07:12:22 crc kubenswrapper[4946]: I1203 07:12:22.100290 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" podUID="af0dc160-2364-4866-b68d-37fe8990241d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Dec 03 07:12:23 crc kubenswrapper[4946]: I1203 07:12:23.039299 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:12:23 crc kubenswrapper[4946]: I1203 07:12:23.039696 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.065390 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.178110 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7740830b-2468-417a-9f4b-6458026070b6","Type":"ContainerDied","Data":"1deffa94aef04dab39663d0ea4d237cde3bc583fa0735ddffa29762c492fec13"} Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.178165 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.241410 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-logs\") pod \"7740830b-2468-417a-9f4b-6458026070b6\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.241491 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-config-data\") pod \"7740830b-2468-417a-9f4b-6458026070b6\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.241527 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-combined-ca-bundle\") pod \"7740830b-2468-417a-9f4b-6458026070b6\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.241597 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcpgp\" (UniqueName: \"kubernetes.io/projected/7740830b-2468-417a-9f4b-6458026070b6-kube-api-access-gcpgp\") pod \"7740830b-2468-417a-9f4b-6458026070b6\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.241646 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-httpd-run\") pod \"7740830b-2468-417a-9f4b-6458026070b6\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.241756 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-scripts\") pod \"7740830b-2468-417a-9f4b-6458026070b6\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.241815 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7740830b-2468-417a-9f4b-6458026070b6\" (UID: \"7740830b-2468-417a-9f4b-6458026070b6\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.242049 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-logs" (OuterVolumeSpecName: "logs") pod "7740830b-2468-417a-9f4b-6458026070b6" (UID: "7740830b-2468-417a-9f4b-6458026070b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.242124 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7740830b-2468-417a-9f4b-6458026070b6" (UID: "7740830b-2468-417a-9f4b-6458026070b6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.243262 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.243283 4946 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7740830b-2468-417a-9f4b-6458026070b6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.247051 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "7740830b-2468-417a-9f4b-6458026070b6" (UID: "7740830b-2468-417a-9f4b-6458026070b6"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.248560 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-scripts" (OuterVolumeSpecName: "scripts") pod "7740830b-2468-417a-9f4b-6458026070b6" (UID: "7740830b-2468-417a-9f4b-6458026070b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.250696 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7740830b-2468-417a-9f4b-6458026070b6-kube-api-access-gcpgp" (OuterVolumeSpecName: "kube-api-access-gcpgp") pod "7740830b-2468-417a-9f4b-6458026070b6" (UID: "7740830b-2468-417a-9f4b-6458026070b6"). InnerVolumeSpecName "kube-api-access-gcpgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.272983 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7740830b-2468-417a-9f4b-6458026070b6" (UID: "7740830b-2468-417a-9f4b-6458026070b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.302254 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-config-data" (OuterVolumeSpecName: "config-data") pod "7740830b-2468-417a-9f4b-6458026070b6" (UID: "7740830b-2468-417a-9f4b-6458026070b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.344128 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.344188 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.344206 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.344218 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7740830b-2468-417a-9f4b-6458026070b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.344231 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcpgp\" (UniqueName: \"kubernetes.io/projected/7740830b-2468-417a-9f4b-6458026070b6-kube-api-access-gcpgp\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.361540 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 07:12:25 crc kubenswrapper[4946]: E1203 07:12:25.439128 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31" Dec 03 07:12:25 crc kubenswrapper[4946]: E1203 07:12:25.439294 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68h679h67h579hc4hb5h54h566h5cch87h54fhfdhcfh557h94h86h56h5c5h585h5cbh6ch595h594hc5h5d4h5ddh677h56ch9h589h77hfbq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xm8pc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(4e32c0a4-449e-4f0f-9ab4-d020d313a99b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.445134 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.456958 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.525623 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.540270 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.556547 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:25 crc kubenswrapper[4946]: E1203 07:12:25.556935 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0dc160-2364-4866-b68d-37fe8990241d" containerName="init" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.556954 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0dc160-2364-4866-b68d-37fe8990241d" containerName="init" Dec 03 07:12:25 crc kubenswrapper[4946]: E1203 07:12:25.556986 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0dc160-2364-4866-b68d-37fe8990241d" containerName="dnsmasq-dns" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.556995 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0dc160-2364-4866-b68d-37fe8990241d" containerName="dnsmasq-dns" Dec 03 07:12:25 crc kubenswrapper[4946]: E1203 07:12:25.557015 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7740830b-2468-417a-9f4b-6458026070b6" containerName="glance-log" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.557022 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7740830b-2468-417a-9f4b-6458026070b6" containerName="glance-log" Dec 03 07:12:25 crc kubenswrapper[4946]: E1203 07:12:25.557041 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7740830b-2468-417a-9f4b-6458026070b6" containerName="glance-httpd" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.557049 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7740830b-2468-417a-9f4b-6458026070b6" containerName="glance-httpd" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.557253 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0dc160-2364-4866-b68d-37fe8990241d" containerName="dnsmasq-dns" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.557289 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7740830b-2468-417a-9f4b-6458026070b6" containerName="glance-httpd" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.557311 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7740830b-2468-417a-9f4b-6458026070b6" containerName="glance-log" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.558566 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.561599 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.564070 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.568718 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.603912 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7740830b-2468-417a-9f4b-6458026070b6" path="/var/lib/kubelet/pods/7740830b-2468-417a-9f4b-6458026070b6/volumes" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.651179 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-sb\") pod \"af0dc160-2364-4866-b68d-37fe8990241d\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.651242 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-nb\") pod \"af0dc160-2364-4866-b68d-37fe8990241d\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.651272 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-config\") pod \"af0dc160-2364-4866-b68d-37fe8990241d\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.651299 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-swift-storage-0\") pod \"af0dc160-2364-4866-b68d-37fe8990241d\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.651404 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cddqw\" (UniqueName: \"kubernetes.io/projected/af0dc160-2364-4866-b68d-37fe8990241d-kube-api-access-cddqw\") pod \"af0dc160-2364-4866-b68d-37fe8990241d\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.651470 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-svc\") pod \"af0dc160-2364-4866-b68d-37fe8990241d\" (UID: \"af0dc160-2364-4866-b68d-37fe8990241d\") " Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.655889 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0dc160-2364-4866-b68d-37fe8990241d-kube-api-access-cddqw" (OuterVolumeSpecName: "kube-api-access-cddqw") pod "af0dc160-2364-4866-b68d-37fe8990241d" (UID: "af0dc160-2364-4866-b68d-37fe8990241d"). InnerVolumeSpecName "kube-api-access-cddqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.697031 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "af0dc160-2364-4866-b68d-37fe8990241d" (UID: "af0dc160-2364-4866-b68d-37fe8990241d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.701092 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "af0dc160-2364-4866-b68d-37fe8990241d" (UID: "af0dc160-2364-4866-b68d-37fe8990241d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.708042 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "af0dc160-2364-4866-b68d-37fe8990241d" (UID: "af0dc160-2364-4866-b68d-37fe8990241d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.708336 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "af0dc160-2364-4866-b68d-37fe8990241d" (UID: "af0dc160-2364-4866-b68d-37fe8990241d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.715411 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-config" (OuterVolumeSpecName: "config") pod "af0dc160-2364-4866-b68d-37fe8990241d" (UID: "af0dc160-2364-4866-b68d-37fe8990241d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754125 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754168 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754192 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754223 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-logs\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754239 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hb4k\" (UniqueName: \"kubernetes.io/projected/0858d6c7-5e35-4c61-b0fd-70cd101c0498-kube-api-access-8hb4k\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754262 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754377 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754409 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754476 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754486 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754498 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cddqw\" (UniqueName: \"kubernetes.io/projected/af0dc160-2364-4866-b68d-37fe8990241d-kube-api-access-cddqw\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754507 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754515 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.754522 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0dc160-2364-4866-b68d-37fe8990241d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.855766 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.855804 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.855832 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.856334 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-logs\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.856362 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hb4k\" (UniqueName: \"kubernetes.io/projected/0858d6c7-5e35-4c61-b0fd-70cd101c0498-kube-api-access-8hb4k\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.856386 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.856426 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.856448 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.856977 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.857155 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-logs\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.857494 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.861532 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.861901 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.863285 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.867429 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.880216 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hb4k\" (UniqueName: \"kubernetes.io/projected/0858d6c7-5e35-4c61-b0fd-70cd101c0498-kube-api-access-8hb4k\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:25 crc kubenswrapper[4946]: I1203 07:12:25.887632 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.174931 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.191078 4946 generic.go:334] "Generic (PLEG): container finished" podID="6cba79e5-c505-4c72-9609-48838c0f59c5" containerID="8dcad6ca973112b601de433a0add9cb4252be0e542acfd5c99157c13bbe11c5a" exitCode=0 Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.191170 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qzv4q" event={"ID":"6cba79e5-c505-4c72-9609-48838c0f59c5","Type":"ContainerDied","Data":"8dcad6ca973112b601de433a0add9cb4252be0e542acfd5c99157c13bbe11c5a"} Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.194863 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" event={"ID":"af0dc160-2364-4866-b68d-37fe8990241d","Type":"ContainerDied","Data":"da198542aabb891a4656828f4f26c609c79ce1f416150dcf2b06bd9e4f517046"} Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.195005 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.233005 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5cc849d9-2r4ds"] Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.239368 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d5cc849d9-2r4ds"] Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.659448 4946 scope.go:117] "RemoveContainer" containerID="d1d9b316ace5b8774924494a238c210e238707a050346e296ee5a326a63523cc" Dec 03 07:12:26 crc kubenswrapper[4946]: E1203 07:12:26.687842 4946 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2" Dec 03 07:12:26 crc kubenswrapper[4946]: E1203 07:12:26.688048 4946 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8x8g8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-9twd7_openstack(8e368226-3cfe-42a1-b2a3-52a53d227807): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:12:26 crc kubenswrapper[4946]: E1203 07:12:26.689551 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-9twd7" podUID="8e368226-3cfe-42a1-b2a3-52a53d227807" Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.802563 4946 scope.go:117] "RemoveContainer" containerID="4ca4cd59686260af266576a5da095516a30bacdc7a29e37f3588e79be9586818" Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.837253 4946 scope.go:117] "RemoveContainer" containerID="b378856018a35c42b4a69b49659e833c49da4683efdeeb418c5ff78884cf1b5a" Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.875007 4946 scope.go:117] "RemoveContainer" containerID="f7fac8d05109eff647ba6132cda7c2e640a242149c274d302f87f99b27b94641" Dec 03 07:12:26 crc kubenswrapper[4946]: I1203 07:12:26.905709 4946 scope.go:117] "RemoveContainer" containerID="75f1b9fda6f1ad88468e26ab6d8942b9d4779f02eff656444fe9648c85ea6137" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.100891 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7d5cc849d9-2r4ds" podUID="af0dc160-2364-4866-b68d-37fe8990241d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.116462 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rwzxk"] Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.203822 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.215537 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rwzxk" event={"ID":"a0b7bfe9-8f09-44f1-8966-3c1df805df6c","Type":"ContainerStarted","Data":"bdbd6cf85583ef6b1d667f4e398a7746b3527f682dd8edd008bb3beff29af146"} Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.216683 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qhs9v" event={"ID":"0190a925-4703-4ae4-8ab4-150aef815c52","Type":"ContainerStarted","Data":"06396dccbda6437120d3bc95eb233be6e528a4975537c272e4321c4e7fb51398"} Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.219125 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qnf4v" event={"ID":"db4ad7e1-ae04-4f3a-8594-2897d4052e32","Type":"ContainerStarted","Data":"2c6c7a3b7db447b64ab670632d0dd83df47c93d671ed2f91a26e1146fa3b06ad"} Dec 03 07:12:27 crc kubenswrapper[4946]: E1203 07:12:27.230353 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2\\\"\"" pod="openstack/cinder-db-sync-9twd7" podUID="8e368226-3cfe-42a1-b2a3-52a53d227807" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.237447 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qhs9v" podStartSLOduration=2.263963262 podStartE2EDuration="23.237432487s" podCreationTimestamp="2025-12-03 07:12:04 +0000 UTC" firstStartedPulling="2025-12-03 07:12:05.632579256 +0000 UTC m=+1318.429269365" lastFinishedPulling="2025-12-03 07:12:26.606048441 +0000 UTC m=+1339.402738590" observedRunningTime="2025-12-03 07:12:27.233571714 +0000 UTC m=+1340.030261823" watchObservedRunningTime="2025-12-03 07:12:27.237432487 +0000 UTC m=+1340.034122596" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.269251 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-qnf4v" podStartSLOduration=2.291756944 podStartE2EDuration="23.269236906s" podCreationTimestamp="2025-12-03 07:12:04 +0000 UTC" firstStartedPulling="2025-12-03 07:12:05.658012736 +0000 UTC m=+1318.454702845" lastFinishedPulling="2025-12-03 07:12:26.635492658 +0000 UTC m=+1339.432182807" observedRunningTime="2025-12-03 07:12:27.268134887 +0000 UTC m=+1340.064824996" watchObservedRunningTime="2025-12-03 07:12:27.269236906 +0000 UTC m=+1340.065927005" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.343586 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:12:27 crc kubenswrapper[4946]: W1203 07:12:27.480648 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0858d6c7_5e35_4c61_b0fd_70cd101c0498.slice/crio-cacc1f5df15dfb52ff77dd2b14a4908d4354bea6ea2f5f7f8119443a9dd73da2 WatchSource:0}: Error finding container cacc1f5df15dfb52ff77dd2b14a4908d4354bea6ea2f5f7f8119443a9dd73da2: Status 404 returned error can't find the container with id cacc1f5df15dfb52ff77dd2b14a4908d4354bea6ea2f5f7f8119443a9dd73da2 Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.586359 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.634363 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0dc160-2364-4866-b68d-37fe8990241d" path="/var/lib/kubelet/pods/af0dc160-2364-4866-b68d-37fe8990241d/volumes" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.688509 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-combined-ca-bundle\") pod \"6cba79e5-c505-4c72-9609-48838c0f59c5\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.688587 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggr95\" (UniqueName: \"kubernetes.io/projected/6cba79e5-c505-4c72-9609-48838c0f59c5-kube-api-access-ggr95\") pod \"6cba79e5-c505-4c72-9609-48838c0f59c5\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.688629 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-config\") pod \"6cba79e5-c505-4c72-9609-48838c0f59c5\" (UID: \"6cba79e5-c505-4c72-9609-48838c0f59c5\") " Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.710190 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cba79e5-c505-4c72-9609-48838c0f59c5-kube-api-access-ggr95" (OuterVolumeSpecName: "kube-api-access-ggr95") pod "6cba79e5-c505-4c72-9609-48838c0f59c5" (UID: "6cba79e5-c505-4c72-9609-48838c0f59c5"). InnerVolumeSpecName "kube-api-access-ggr95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.722169 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cba79e5-c505-4c72-9609-48838c0f59c5" (UID: "6cba79e5-c505-4c72-9609-48838c0f59c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.740486 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-config" (OuterVolumeSpecName: "config") pod "6cba79e5-c505-4c72-9609-48838c0f59c5" (UID: "6cba79e5-c505-4c72-9609-48838c0f59c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.791793 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.791835 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggr95\" (UniqueName: \"kubernetes.io/projected/6cba79e5-c505-4c72-9609-48838c0f59c5-kube-api-access-ggr95\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:27 crc kubenswrapper[4946]: I1203 07:12:27.791850 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cba79e5-c505-4c72-9609-48838c0f59c5-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.254657 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0858d6c7-5e35-4c61-b0fd-70cd101c0498","Type":"ContainerStarted","Data":"3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517"} Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.254965 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0858d6c7-5e35-4c61-b0fd-70cd101c0498","Type":"ContainerStarted","Data":"cacc1f5df15dfb52ff77dd2b14a4908d4354bea6ea2f5f7f8119443a9dd73da2"} Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.266250 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qzv4q" event={"ID":"6cba79e5-c505-4c72-9609-48838c0f59c5","Type":"ContainerDied","Data":"782ba6b6aebd5af6d1efaa97f6845d2c93e271022e62716c9c6b3e77cc1ec021"} Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.266299 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="782ba6b6aebd5af6d1efaa97f6845d2c93e271022e62716c9c6b3e77cc1ec021" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.266377 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qzv4q" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.299618 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rwzxk" event={"ID":"a0b7bfe9-8f09-44f1-8966-3c1df805df6c","Type":"ContainerStarted","Data":"9cccbf44990a5cf20986b72a9645414dda943c7dde99da340999a5cf27c3cfe5"} Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.307391 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e32c0a4-449e-4f0f-9ab4-d020d313a99b","Type":"ContainerStarted","Data":"61ea719a4977bd323e741afe2c8d71bdfb58887990345513124795cd4a6f53a9"} Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.321511 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7","Type":"ContainerStarted","Data":"6e66bc9d1b8394a530308eae9d3fb3ffc5e8f1512eef972c0d4c3b73a556409d"} Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.321777 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7","Type":"ContainerStarted","Data":"0cc93719d4efd02d1e0636175b9f92adb30f05187af5552a57a9578619eccea9"} Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.343396 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rwzxk" podStartSLOduration=14.343378318 podStartE2EDuration="14.343378318s" podCreationTimestamp="2025-12-03 07:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:28.334720137 +0000 UTC m=+1341.131410246" watchObservedRunningTime="2025-12-03 07:12:28.343378318 +0000 UTC m=+1341.140068427" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.368190 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c874f55b9-9l9h8"] Dec 03 07:12:28 crc kubenswrapper[4946]: E1203 07:12:28.368979 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cba79e5-c505-4c72-9609-48838c0f59c5" containerName="neutron-db-sync" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.368996 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cba79e5-c505-4c72-9609-48838c0f59c5" containerName="neutron-db-sync" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.369254 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cba79e5-c505-4c72-9609-48838c0f59c5" containerName="neutron-db-sync" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.377076 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.405622 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c874f55b9-9l9h8"] Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.483540 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6945f689f6-c4fbq"] Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.485318 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.489516 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.489788 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-q8hq6" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.489922 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.490031 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.499104 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6945f689f6-c4fbq"] Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.515416 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-swift-storage-0\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.515566 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-nb\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.515696 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbnpx\" (UniqueName: \"kubernetes.io/projected/105df626-6fb4-453a-9851-836d6d615e6a-kube-api-access-zbnpx\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.515868 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-svc\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.515939 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-config\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.516066 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-sb\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618657 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-config\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618758 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-ovndb-tls-certs\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618792 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-sb\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618810 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-config\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618837 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-swift-storage-0\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618872 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-nb\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618926 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbnpx\" (UniqueName: \"kubernetes.io/projected/105df626-6fb4-453a-9851-836d6d615e6a-kube-api-access-zbnpx\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618945 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-combined-ca-bundle\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618967 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-httpd-config\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.618985 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-svc\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.619014 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqb7k\" (UniqueName: \"kubernetes.io/projected/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-kube-api-access-sqb7k\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.619967 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-config\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.620161 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-sb\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.621986 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-svc\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.623393 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-swift-storage-0\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.623590 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-nb\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.637611 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbnpx\" (UniqueName: \"kubernetes.io/projected/105df626-6fb4-453a-9851-836d6d615e6a-kube-api-access-zbnpx\") pod \"dnsmasq-dns-7c874f55b9-9l9h8\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.720406 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.720828 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-combined-ca-bundle\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.720864 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-httpd-config\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.720892 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqb7k\" (UniqueName: \"kubernetes.io/projected/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-kube-api-access-sqb7k\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.720954 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-ovndb-tls-certs\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.720989 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-config\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.723981 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-combined-ca-bundle\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.733481 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-config\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.733540 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-ovndb-tls-certs\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.734115 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-httpd-config\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.736715 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqb7k\" (UniqueName: \"kubernetes.io/projected/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-kube-api-access-sqb7k\") pod \"neutron-6945f689f6-c4fbq\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:28 crc kubenswrapper[4946]: I1203 07:12:28.808479 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:29 crc kubenswrapper[4946]: I1203 07:12:29.333941 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7","Type":"ContainerStarted","Data":"7664bde7ec828564ba96e4d3d5111e746146cbd7db9fad042b1af00c47996a16"} Dec 03 07:12:29 crc kubenswrapper[4946]: I1203 07:12:29.334701 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c874f55b9-9l9h8"] Dec 03 07:12:29 crc kubenswrapper[4946]: I1203 07:12:29.339214 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0858d6c7-5e35-4c61-b0fd-70cd101c0498","Type":"ContainerStarted","Data":"3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1"} Dec 03 07:12:29 crc kubenswrapper[4946]: I1203 07:12:29.374345 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=13.374323756999999 podStartE2EDuration="13.374323757s" podCreationTimestamp="2025-12-03 07:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:29.361940977 +0000 UTC m=+1342.158631086" watchObservedRunningTime="2025-12-03 07:12:29.374323757 +0000 UTC m=+1342.171013866" Dec 03 07:12:29 crc kubenswrapper[4946]: I1203 07:12:29.391983 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.391964809 podStartE2EDuration="4.391964809s" podCreationTimestamp="2025-12-03 07:12:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:29.389124113 +0000 UTC m=+1342.185814232" watchObservedRunningTime="2025-12-03 07:12:29.391964809 +0000 UTC m=+1342.188654918" Dec 03 07:12:29 crc kubenswrapper[4946]: I1203 07:12:29.463209 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6945f689f6-c4fbq"] Dec 03 07:12:29 crc kubenswrapper[4946]: W1203 07:12:29.469282 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c80a231_30ed_4e17_b8ba_20b628b8b9bd.slice/crio-687090c41ce1ed4585057f91a2df313a47f4caef1c76cfa8f354d076c8a7deda WatchSource:0}: Error finding container 687090c41ce1ed4585057f91a2df313a47f4caef1c76cfa8f354d076c8a7deda: Status 404 returned error can't find the container with id 687090c41ce1ed4585057f91a2df313a47f4caef1c76cfa8f354d076c8a7deda Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.347174 4946 generic.go:334] "Generic (PLEG): container finished" podID="105df626-6fb4-453a-9851-836d6d615e6a" containerID="52bf577729b1d37071e43ab787e4067a963215f2de30d155886f5584744298da" exitCode=0 Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.347258 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" event={"ID":"105df626-6fb4-453a-9851-836d6d615e6a","Type":"ContainerDied","Data":"52bf577729b1d37071e43ab787e4067a963215f2de30d155886f5584744298da"} Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.347684 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" event={"ID":"105df626-6fb4-453a-9851-836d6d615e6a","Type":"ContainerStarted","Data":"24926b6051fbebea69be509c1d7f3f172adb5b85657b3193672b8e0babec6719"} Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.352215 4946 generic.go:334] "Generic (PLEG): container finished" podID="0190a925-4703-4ae4-8ab4-150aef815c52" containerID="06396dccbda6437120d3bc95eb233be6e528a4975537c272e4321c4e7fb51398" exitCode=0 Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.352281 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qhs9v" event={"ID":"0190a925-4703-4ae4-8ab4-150aef815c52","Type":"ContainerDied","Data":"06396dccbda6437120d3bc95eb233be6e528a4975537c272e4321c4e7fb51398"} Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.358860 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6945f689f6-c4fbq" event={"ID":"9c80a231-30ed-4e17-b8ba-20b628b8b9bd","Type":"ContainerStarted","Data":"7e5a02adf9b400b9fc12a49adbdc155b358023367bd475eefec73c51efc530e7"} Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.358886 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6945f689f6-c4fbq" event={"ID":"9c80a231-30ed-4e17-b8ba-20b628b8b9bd","Type":"ContainerStarted","Data":"bbf789b9f76a02d9d9c15d2b2bb403dd915d29bc873e449ba1e2b5fde5777c80"} Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.358895 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6945f689f6-c4fbq" event={"ID":"9c80a231-30ed-4e17-b8ba-20b628b8b9bd","Type":"ContainerStarted","Data":"687090c41ce1ed4585057f91a2df313a47f4caef1c76cfa8f354d076c8a7deda"} Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.359538 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:30 crc kubenswrapper[4946]: I1203 07:12:30.418373 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6945f689f6-c4fbq" podStartSLOduration=2.418356196 podStartE2EDuration="2.418356196s" podCreationTimestamp="2025-12-03 07:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:30.384948234 +0000 UTC m=+1343.181638353" watchObservedRunningTime="2025-12-03 07:12:30.418356196 +0000 UTC m=+1343.215046305" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.342851 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-86855dfc4c-jdcjc"] Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.344550 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.350208 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.351800 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.362857 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86855dfc4c-jdcjc"] Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.368327 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-ovndb-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.368390 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-public-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.368452 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-httpd-config\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.368487 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sls7d\" (UniqueName: \"kubernetes.io/projected/cc385f93-2111-42cf-93be-d4cc82225570-kube-api-access-sls7d\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.368551 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-combined-ca-bundle\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.368597 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-internal-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.368616 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-config\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.381974 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" event={"ID":"105df626-6fb4-453a-9851-836d6d615e6a","Type":"ContainerStarted","Data":"680dbd6776ef738d33ff6b66495a39019dd36451bb0ed629ebce9c55a8ca06b9"} Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.383328 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.397703 4946 generic.go:334] "Generic (PLEG): container finished" podID="db4ad7e1-ae04-4f3a-8594-2897d4052e32" containerID="2c6c7a3b7db447b64ab670632d0dd83df47c93d671ed2f91a26e1146fa3b06ad" exitCode=0 Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.400026 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qnf4v" event={"ID":"db4ad7e1-ae04-4f3a-8594-2897d4052e32","Type":"ContainerDied","Data":"2c6c7a3b7db447b64ab670632d0dd83df47c93d671ed2f91a26e1146fa3b06ad"} Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.408897 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" podStartSLOduration=3.408873634 podStartE2EDuration="3.408873634s" podCreationTimestamp="2025-12-03 07:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:31.405802042 +0000 UTC m=+1344.202492151" watchObservedRunningTime="2025-12-03 07:12:31.408873634 +0000 UTC m=+1344.205563733" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.471900 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sls7d\" (UniqueName: \"kubernetes.io/projected/cc385f93-2111-42cf-93be-d4cc82225570-kube-api-access-sls7d\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.472094 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-combined-ca-bundle\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.472168 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-internal-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.472194 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-config\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.472796 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-ovndb-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.472861 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-public-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.472920 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-httpd-config\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.477576 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-public-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.478305 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-combined-ca-bundle\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.479028 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-config\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.480538 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-ovndb-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.494411 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sls7d\" (UniqueName: \"kubernetes.io/projected/cc385f93-2111-42cf-93be-d4cc82225570-kube-api-access-sls7d\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.498186 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-internal-tls-certs\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.498570 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-httpd-config\") pod \"neutron-86855dfc4c-jdcjc\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:31 crc kubenswrapper[4946]: I1203 07:12:31.663456 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:32 crc kubenswrapper[4946]: I1203 07:12:32.436335 4946 generic.go:334] "Generic (PLEG): container finished" podID="a0b7bfe9-8f09-44f1-8966-3c1df805df6c" containerID="9cccbf44990a5cf20986b72a9645414dda943c7dde99da340999a5cf27c3cfe5" exitCode=0 Dec 03 07:12:32 crc kubenswrapper[4946]: I1203 07:12:32.436983 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rwzxk" event={"ID":"a0b7bfe9-8f09-44f1-8966-3c1df805df6c","Type":"ContainerDied","Data":"9cccbf44990a5cf20986b72a9645414dda943c7dde99da340999a5cf27c3cfe5"} Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.479683 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rwzxk" event={"ID":"a0b7bfe9-8f09-44f1-8966-3c1df805df6c","Type":"ContainerDied","Data":"bdbd6cf85583ef6b1d667f4e398a7746b3527f682dd8edd008bb3beff29af146"} Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.480338 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdbd6cf85583ef6b1d667f4e398a7746b3527f682dd8edd008bb3beff29af146" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.496206 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qhs9v" event={"ID":"0190a925-4703-4ae4-8ab4-150aef815c52","Type":"ContainerDied","Data":"c7736f10fd024ff8119cc401925d35c4d22d580e40c4e2149d932f9904fb136d"} Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.496241 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7736f10fd024ff8119cc401925d35c4d22d580e40c4e2149d932f9904fb136d" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.503422 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.504642 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qnf4v" event={"ID":"db4ad7e1-ae04-4f3a-8594-2897d4052e32","Type":"ContainerDied","Data":"90f69fd381331cc338f5a8564f1d853a69322ce9f61fc3cc834d37b13077a628"} Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.504679 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90f69fd381331cc338f5a8564f1d853a69322ce9f61fc3cc834d37b13077a628" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.526783 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.550989 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647006 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwmw2\" (UniqueName: \"kubernetes.io/projected/0190a925-4703-4ae4-8ab4-150aef815c52-kube-api-access-fwmw2\") pod \"0190a925-4703-4ae4-8ab4-150aef815c52\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647274 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0190a925-4703-4ae4-8ab4-150aef815c52-logs\") pod \"0190a925-4703-4ae4-8ab4-150aef815c52\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647317 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-config-data\") pod \"0190a925-4703-4ae4-8ab4-150aef815c52\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647387 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-combined-ca-bundle\") pod \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647435 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-db-sync-config-data\") pod \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647821 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x44l2\" (UniqueName: \"kubernetes.io/projected/db4ad7e1-ae04-4f3a-8594-2897d4052e32-kube-api-access-x44l2\") pod \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\" (UID: \"db4ad7e1-ae04-4f3a-8594-2897d4052e32\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647863 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-scripts\") pod \"0190a925-4703-4ae4-8ab4-150aef815c52\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647882 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-combined-ca-bundle\") pod \"0190a925-4703-4ae4-8ab4-150aef815c52\" (UID: \"0190a925-4703-4ae4-8ab4-150aef815c52\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.647946 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s2g2\" (UniqueName: \"kubernetes.io/projected/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-kube-api-access-4s2g2\") pod \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.648416 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0190a925-4703-4ae4-8ab4-150aef815c52-logs" (OuterVolumeSpecName: "logs") pod "0190a925-4703-4ae4-8ab4-150aef815c52" (UID: "0190a925-4703-4ae4-8ab4-150aef815c52"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.651790 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "db4ad7e1-ae04-4f3a-8594-2897d4052e32" (UID: "db4ad7e1-ae04-4f3a-8594-2897d4052e32"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.652281 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-scripts" (OuterVolumeSpecName: "scripts") pod "0190a925-4703-4ae4-8ab4-150aef815c52" (UID: "0190a925-4703-4ae4-8ab4-150aef815c52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.652285 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4ad7e1-ae04-4f3a-8594-2897d4052e32-kube-api-access-x44l2" (OuterVolumeSpecName: "kube-api-access-x44l2") pod "db4ad7e1-ae04-4f3a-8594-2897d4052e32" (UID: "db4ad7e1-ae04-4f3a-8594-2897d4052e32"). InnerVolumeSpecName "kube-api-access-x44l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.652702 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-kube-api-access-4s2g2" (OuterVolumeSpecName: "kube-api-access-4s2g2") pod "a0b7bfe9-8f09-44f1-8966-3c1df805df6c" (UID: "a0b7bfe9-8f09-44f1-8966-3c1df805df6c"). InnerVolumeSpecName "kube-api-access-4s2g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.668183 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0190a925-4703-4ae4-8ab4-150aef815c52-kube-api-access-fwmw2" (OuterVolumeSpecName: "kube-api-access-fwmw2") pod "0190a925-4703-4ae4-8ab4-150aef815c52" (UID: "0190a925-4703-4ae4-8ab4-150aef815c52"). InnerVolumeSpecName "kube-api-access-fwmw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.671560 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0190a925-4703-4ae4-8ab4-150aef815c52" (UID: "0190a925-4703-4ae4-8ab4-150aef815c52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.673005 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-config-data" (OuterVolumeSpecName: "config-data") pod "0190a925-4703-4ae4-8ab4-150aef815c52" (UID: "0190a925-4703-4ae4-8ab4-150aef815c52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.682807 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db4ad7e1-ae04-4f3a-8594-2897d4052e32" (UID: "db4ad7e1-ae04-4f3a-8594-2897d4052e32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.749268 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-scripts\") pod \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.749493 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-credential-keys\") pod \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.749543 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-config-data\") pod \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.749661 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-fernet-keys\") pod \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.749705 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-combined-ca-bundle\") pod \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\" (UID: \"a0b7bfe9-8f09-44f1-8966-3c1df805df6c\") " Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750183 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750207 4946 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/db4ad7e1-ae04-4f3a-8594-2897d4052e32-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750221 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x44l2\" (UniqueName: \"kubernetes.io/projected/db4ad7e1-ae04-4f3a-8594-2897d4052e32-kube-api-access-x44l2\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750234 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750246 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750258 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s2g2\" (UniqueName: \"kubernetes.io/projected/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-kube-api-access-4s2g2\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750270 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwmw2\" (UniqueName: \"kubernetes.io/projected/0190a925-4703-4ae4-8ab4-150aef815c52-kube-api-access-fwmw2\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750283 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0190a925-4703-4ae4-8ab4-150aef815c52-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.750294 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0190a925-4703-4ae4-8ab4-150aef815c52-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.753100 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a0b7bfe9-8f09-44f1-8966-3c1df805df6c" (UID: "a0b7bfe9-8f09-44f1-8966-3c1df805df6c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.754472 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a0b7bfe9-8f09-44f1-8966-3c1df805df6c" (UID: "a0b7bfe9-8f09-44f1-8966-3c1df805df6c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.754955 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-scripts" (OuterVolumeSpecName: "scripts") pod "a0b7bfe9-8f09-44f1-8966-3c1df805df6c" (UID: "a0b7bfe9-8f09-44f1-8966-3c1df805df6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.775402 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-config-data" (OuterVolumeSpecName: "config-data") pod "a0b7bfe9-8f09-44f1-8966-3c1df805df6c" (UID: "a0b7bfe9-8f09-44f1-8966-3c1df805df6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.780807 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0b7bfe9-8f09-44f1-8966-3c1df805df6c" (UID: "a0b7bfe9-8f09-44f1-8966-3c1df805df6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.851923 4946 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.851990 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.852016 4946 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.852040 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.852064 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b7bfe9-8f09-44f1-8966-3c1df805df6c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:35 crc kubenswrapper[4946]: I1203 07:12:35.883625 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86855dfc4c-jdcjc"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.175252 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.176369 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.227381 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.242326 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.493821 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.493885 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.518701 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86855dfc4c-jdcjc" event={"ID":"cc385f93-2111-42cf-93be-d4cc82225570","Type":"ContainerStarted","Data":"01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1"} Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.518767 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86855dfc4c-jdcjc" event={"ID":"cc385f93-2111-42cf-93be-d4cc82225570","Type":"ContainerStarted","Data":"a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be"} Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.518785 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86855dfc4c-jdcjc" event={"ID":"cc385f93-2111-42cf-93be-d4cc82225570","Type":"ContainerStarted","Data":"dc75f61ddd5966e8f97d174ed529f5c861cb0f2d892d432e30b91814e9ca814e"} Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.519865 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.523005 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qnf4v" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.523051 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qhs9v" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.523081 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rwzxk" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.537368 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e32c0a4-449e-4f0f-9ab4-d020d313a99b","Type":"ContainerStarted","Data":"c55e987e9d86be63d4718b1bc16fe9a74d06571d798460e0883408639260cec0"} Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.537444 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.537462 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.552111 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.552882 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.583720 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-86855dfc4c-jdcjc" podStartSLOduration=5.583702065 podStartE2EDuration="5.583702065s" podCreationTimestamp="2025-12-03 07:12:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:36.580015996 +0000 UTC m=+1349.376706115" watchObservedRunningTime="2025-12-03 07:12:36.583702065 +0000 UTC m=+1349.380392174" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.586510 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.624240 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-69454fc688-t54w7"] Dec 03 07:12:36 crc kubenswrapper[4946]: E1203 07:12:36.624855 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4ad7e1-ae04-4f3a-8594-2897d4052e32" containerName="barbican-db-sync" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.624871 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4ad7e1-ae04-4f3a-8594-2897d4052e32" containerName="barbican-db-sync" Dec 03 07:12:36 crc kubenswrapper[4946]: E1203 07:12:36.625013 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0190a925-4703-4ae4-8ab4-150aef815c52" containerName="placement-db-sync" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.625026 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0190a925-4703-4ae4-8ab4-150aef815c52" containerName="placement-db-sync" Dec 03 07:12:36 crc kubenswrapper[4946]: E1203 07:12:36.625041 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b7bfe9-8f09-44f1-8966-3c1df805df6c" containerName="keystone-bootstrap" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.625048 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b7bfe9-8f09-44f1-8966-3c1df805df6c" containerName="keystone-bootstrap" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.625440 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b7bfe9-8f09-44f1-8966-3c1df805df6c" containerName="keystone-bootstrap" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.625458 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4ad7e1-ae04-4f3a-8594-2897d4052e32" containerName="barbican-db-sync" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.625470 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0190a925-4703-4ae4-8ab4-150aef815c52" containerName="placement-db-sync" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.628574 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.632362 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.633169 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.633381 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.633630 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.633800 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xlgln" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.659247 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69454fc688-t54w7"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.742707 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6b848b8657-gfw2m"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.743938 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.747293 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.747453 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.747551 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.747659 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mwdjs" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.747788 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.747959 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.753922 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6b848b8657-gfw2m"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.777135 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-public-tls-certs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.777187 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-logs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.777249 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-internal-tls-certs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.777282 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-scripts\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.777306 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-combined-ca-bundle\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.777335 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp6zn\" (UniqueName: \"kubernetes.io/projected/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-kube-api-access-zp6zn\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.777369 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-config-data\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.837693 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7f55c455cd-jdhgw"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.839193 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.844994 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.845133 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.848194 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f55c455cd-jdhgw"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.851214 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rkhgb" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.869044 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-559fcd84cc-lhklt"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.870553 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.875019 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.877249 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-559fcd84cc-lhklt"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.884608 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-config-data\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886192 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-public-tls-certs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886503 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-config-data\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886548 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-logs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886583 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw6qg\" (UniqueName: \"kubernetes.io/projected/e30f3ad7-043b-4212-a496-656c3447b509-kube-api-access-bw6qg\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886648 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-internal-tls-certs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886678 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-scripts\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886693 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-internal-tls-certs\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886859 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-scripts\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886896 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-public-tls-certs\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886925 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-combined-ca-bundle\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.886969 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp6zn\" (UniqueName: \"kubernetes.io/projected/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-kube-api-access-zp6zn\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.887013 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-fernet-keys\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.887031 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-combined-ca-bundle\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.887046 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-credential-keys\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.887334 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-logs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.892079 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-combined-ca-bundle\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.893170 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-config-data\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.901069 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-scripts\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.902426 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-public-tls-certs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.906690 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-internal-tls-certs\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.912277 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp6zn\" (UniqueName: \"kubernetes.io/projected/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-kube-api-access-zp6zn\") pod \"placement-69454fc688-t54w7\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.933619 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c874f55b9-9l9h8"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.933878 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" podUID="105df626-6fb4-453a-9851-836d6d615e6a" containerName="dnsmasq-dns" containerID="cri-o://680dbd6776ef738d33ff6b66495a39019dd36451bb0ed629ebce9c55a8ca06b9" gracePeriod=10 Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.934942 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.953782 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.963214 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f568c56f7-69bfr"] Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.965044 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988487 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988541 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-fernet-keys\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988559 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-combined-ca-bundle\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988575 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-credential-keys\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988614 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-config-data\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988644 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ac177b-b889-43e7-888e-8ed02fb3ac72-logs\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988673 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data-custom\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988698 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2g8l\" (UniqueName: \"kubernetes.io/projected/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-kube-api-access-h2g8l\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988723 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw6qg\" (UniqueName: \"kubernetes.io/projected/e30f3ad7-043b-4212-a496-656c3447b509-kube-api-access-bw6qg\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988772 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-combined-ca-bundle\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988796 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data-custom\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988829 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-scripts\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988850 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-internal-tls-certs\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988869 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988885 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-logs\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988909 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-combined-ca-bundle\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988927 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzd5s\" (UniqueName: \"kubernetes.io/projected/26ac177b-b889-43e7-888e-8ed02fb3ac72-kube-api-access-kzd5s\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:36 crc kubenswrapper[4946]: I1203 07:12:36.988959 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-public-tls-certs\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.000165 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-scripts\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.002543 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-fernet-keys\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.013578 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-config-data\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.014360 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-public-tls-certs\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.014613 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-internal-tls-certs\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.016820 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-credential-keys\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.032360 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw6qg\" (UniqueName: \"kubernetes.io/projected/e30f3ad7-043b-4212-a496-656c3447b509-kube-api-access-bw6qg\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.055212 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-combined-ca-bundle\") pod \"keystone-6b848b8657-gfw2m\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.069130 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f568c56f7-69bfr"] Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.081122 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.093950 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-svc\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094061 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-nb\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094177 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094300 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-swift-storage-0\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094393 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ac177b-b889-43e7-888e-8ed02fb3ac72-logs\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094459 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-sb\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094521 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data-custom\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094590 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2g8l\" (UniqueName: \"kubernetes.io/projected/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-kube-api-access-h2g8l\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094666 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-combined-ca-bundle\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094730 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jxfz\" (UniqueName: \"kubernetes.io/projected/64332a7d-18bc-45c6-aad3-86c19e2979c4-kube-api-access-5jxfz\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094817 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data-custom\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094897 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-config\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.094982 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.095050 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-logs\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.095117 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-combined-ca-bundle\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.095181 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzd5s\" (UniqueName: \"kubernetes.io/projected/26ac177b-b889-43e7-888e-8ed02fb3ac72-kube-api-access-kzd5s\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.097671 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-logs\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.098708 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ac177b-b889-43e7-888e-8ed02fb3ac72-logs\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.103110 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.103185 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.104976 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-combined-ca-bundle\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.109543 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-combined-ca-bundle\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.111119 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data-custom\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.119566 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data-custom\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.136243 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2g8l\" (UniqueName: \"kubernetes.io/projected/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-kube-api-access-h2g8l\") pod \"barbican-keystone-listener-7f55c455cd-jdhgw\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.136562 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzd5s\" (UniqueName: \"kubernetes.io/projected/26ac177b-b889-43e7-888e-8ed02fb3ac72-kube-api-access-kzd5s\") pod \"barbican-worker-559fcd84cc-lhklt\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.157137 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.163358 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.180514 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-76b8c8854b-nphm5"] Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.182548 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.185640 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.193263 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-76b8c8854b-nphm5"] Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.200025 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-swift-storage-0\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.200331 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-sb\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.200440 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jxfz\" (UniqueName: \"kubernetes.io/projected/64332a7d-18bc-45c6-aad3-86c19e2979c4-kube-api-access-5jxfz\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.200521 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-config\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.200633 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-svc\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.200704 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-nb\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.202212 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-nb\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.205142 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-swift-storage-0\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.205675 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-sb\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.206459 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-config\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.207003 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-svc\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.227502 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jxfz\" (UniqueName: \"kubernetes.io/projected/64332a7d-18bc-45c6-aad3-86c19e2979c4-kube-api-access-5jxfz\") pod \"dnsmasq-dns-6f568c56f7-69bfr\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.316836 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-combined-ca-bundle\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.318456 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data-custom\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.318494 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cf777b2-9dab-49de-9f2c-a119853437b3-logs\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.318606 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.318670 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxr4t\" (UniqueName: \"kubernetes.io/projected/0cf777b2-9dab-49de-9f2c-a119853437b3-kube-api-access-lxr4t\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.420975 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.421027 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxr4t\" (UniqueName: \"kubernetes.io/projected/0cf777b2-9dab-49de-9f2c-a119853437b3-kube-api-access-lxr4t\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.421069 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-combined-ca-bundle\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.421102 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data-custom\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.421123 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cf777b2-9dab-49de-9f2c-a119853437b3-logs\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.421568 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cf777b2-9dab-49de-9f2c-a119853437b3-logs\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.436412 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data-custom\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.440900 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-combined-ca-bundle\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.442402 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.450232 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxr4t\" (UniqueName: \"kubernetes.io/projected/0cf777b2-9dab-49de-9f2c-a119853437b3-kube-api-access-lxr4t\") pod \"barbican-api-76b8c8854b-nphm5\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.475841 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.513762 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.561046 4946 generic.go:334] "Generic (PLEG): container finished" podID="105df626-6fb4-453a-9851-836d6d615e6a" containerID="680dbd6776ef738d33ff6b66495a39019dd36451bb0ed629ebce9c55a8ca06b9" exitCode=0 Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.561674 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" event={"ID":"105df626-6fb4-453a-9851-836d6d615e6a","Type":"ContainerDied","Data":"680dbd6776ef738d33ff6b66495a39019dd36451bb0ed629ebce9c55a8ca06b9"} Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.564582 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.828939 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69454fc688-t54w7"] Dec 03 07:12:37 crc kubenswrapper[4946]: I1203 07:12:37.905864 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.032020 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-config\") pod \"105df626-6fb4-453a-9851-836d6d615e6a\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.032375 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-swift-storage-0\") pod \"105df626-6fb4-453a-9851-836d6d615e6a\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.032419 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-sb\") pod \"105df626-6fb4-453a-9851-836d6d615e6a\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.032541 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-svc\") pod \"105df626-6fb4-453a-9851-836d6d615e6a\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.032588 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-nb\") pod \"105df626-6fb4-453a-9851-836d6d615e6a\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.034025 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbnpx\" (UniqueName: \"kubernetes.io/projected/105df626-6fb4-453a-9851-836d6d615e6a-kube-api-access-zbnpx\") pod \"105df626-6fb4-453a-9851-836d6d615e6a\" (UID: \"105df626-6fb4-453a-9851-836d6d615e6a\") " Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.039758 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/105df626-6fb4-453a-9851-836d6d615e6a-kube-api-access-zbnpx" (OuterVolumeSpecName: "kube-api-access-zbnpx") pod "105df626-6fb4-453a-9851-836d6d615e6a" (UID: "105df626-6fb4-453a-9851-836d6d615e6a"). InnerVolumeSpecName "kube-api-access-zbnpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.134685 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "105df626-6fb4-453a-9851-836d6d615e6a" (UID: "105df626-6fb4-453a-9851-836d6d615e6a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.144887 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbnpx\" (UniqueName: \"kubernetes.io/projected/105df626-6fb4-453a-9851-836d6d615e6a-kube-api-access-zbnpx\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.144921 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.194925 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "105df626-6fb4-453a-9851-836d6d615e6a" (UID: "105df626-6fb4-453a-9851-836d6d615e6a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.218716 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "105df626-6fb4-453a-9851-836d6d615e6a" (UID: "105df626-6fb4-453a-9851-836d6d615e6a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.237031 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-559fcd84cc-lhklt"] Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.247847 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.247881 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.268776 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6b848b8657-gfw2m"] Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.294761 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f55c455cd-jdhgw"] Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.366576 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f568c56f7-69bfr"] Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.376588 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-76b8c8854b-nphm5"] Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.379774 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-config" (OuterVolumeSpecName: "config") pod "105df626-6fb4-453a-9851-836d6d615e6a" (UID: "105df626-6fb4-453a-9851-836d6d615e6a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:38 crc kubenswrapper[4946]: W1203 07:12:38.400175 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64332a7d_18bc_45c6_aad3_86c19e2979c4.slice/crio-0bd8bbc572e17859fac80900b57065a3dcf824272f38b7df050320fd21279e40 WatchSource:0}: Error finding container 0bd8bbc572e17859fac80900b57065a3dcf824272f38b7df050320fd21279e40: Status 404 returned error can't find the container with id 0bd8bbc572e17859fac80900b57065a3dcf824272f38b7df050320fd21279e40 Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.400553 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "105df626-6fb4-453a-9851-836d6d615e6a" (UID: "105df626-6fb4-453a-9851-836d6d615e6a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.452835 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.453188 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105df626-6fb4-453a-9851-836d6d615e6a-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.607944 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" event={"ID":"105df626-6fb4-453a-9851-836d6d615e6a","Type":"ContainerDied","Data":"24926b6051fbebea69be509c1d7f3f172adb5b85657b3193672b8e0babec6719"} Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.608012 4946 scope.go:117] "RemoveContainer" containerID="680dbd6776ef738d33ff6b66495a39019dd36451bb0ed629ebce9c55a8ca06b9" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.607960 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c874f55b9-9l9h8" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.613074 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b848b8657-gfw2m" event={"ID":"e30f3ad7-043b-4212-a496-656c3447b509","Type":"ContainerStarted","Data":"f14ba74c11fbec5585472424d5c2360f65091f851ecf574c28a4a6adc40dd3aa"} Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.619231 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69454fc688-t54w7" event={"ID":"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541","Type":"ContainerStarted","Data":"f0dbde89ae62dd0d52329bd682bf4a8a258dd6f7c276606d012df0778913d590"} Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.621402 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-559fcd84cc-lhklt" event={"ID":"26ac177b-b889-43e7-888e-8ed02fb3ac72","Type":"ContainerStarted","Data":"50cd82febee37545cca600c5780372aa7dfb6ae8d5ff07654a79fd9977964e38"} Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.627963 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76b8c8854b-nphm5" event={"ID":"0cf777b2-9dab-49de-9f2c-a119853437b3","Type":"ContainerStarted","Data":"473d31036a45bddce221695b524d1dc6fe3ed1343117b99b649f7a6754c81e8d"} Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.631969 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" event={"ID":"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e","Type":"ContainerStarted","Data":"068f56e9cdc616e48e6fcc81467f0f7484fea27c9e42d09a6874ef75ea9bbcd8"} Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.635700 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.635803 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" event={"ID":"64332a7d-18bc-45c6-aad3-86c19e2979c4","Type":"ContainerStarted","Data":"0bd8bbc572e17859fac80900b57065a3dcf824272f38b7df050320fd21279e40"} Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.635899 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.635912 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.655015 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c874f55b9-9l9h8"] Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.668967 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c874f55b9-9l9h8"] Dec 03 07:12:38 crc kubenswrapper[4946]: I1203 07:12:38.855871 4946 scope.go:117] "RemoveContainer" containerID="52bf577729b1d37071e43ab787e4067a963215f2de30d155886f5584744298da" Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.607683 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="105df626-6fb4-453a-9851-836d6d615e6a" path="/var/lib/kubelet/pods/105df626-6fb4-453a-9851-836d6d615e6a/volumes" Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.660677 4946 generic.go:334] "Generic (PLEG): container finished" podID="64332a7d-18bc-45c6-aad3-86c19e2979c4" containerID="362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a" exitCode=0 Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.660765 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" event={"ID":"64332a7d-18bc-45c6-aad3-86c19e2979c4","Type":"ContainerDied","Data":"362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a"} Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.676333 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b848b8657-gfw2m" event={"ID":"e30f3ad7-043b-4212-a496-656c3447b509","Type":"ContainerStarted","Data":"6f451caacded099e931e441c82a111b08382c08a1a8b89ed90e33967d206ea41"} Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.676694 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.691939 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69454fc688-t54w7" event={"ID":"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541","Type":"ContainerStarted","Data":"6a378fe14ccec4bc8e63806ae734f9293c3f59cfa0f3b3d5fb967e83eb1a42a1"} Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.704770 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76b8c8854b-nphm5" event={"ID":"0cf777b2-9dab-49de-9f2c-a119853437b3","Type":"ContainerStarted","Data":"a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f"} Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.710219 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6b848b8657-gfw2m" podStartSLOduration=3.71019574 podStartE2EDuration="3.71019574s" podCreationTimestamp="2025-12-03 07:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:39.70197284 +0000 UTC m=+1352.498662949" watchObservedRunningTime="2025-12-03 07:12:39.71019574 +0000 UTC m=+1352.506885859" Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.811362 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.811633 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.956636 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 07:12:39 crc kubenswrapper[4946]: I1203 07:12:39.956720 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.007285 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.428809 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6786b69d4-clggf"] Dec 03 07:12:40 crc kubenswrapper[4946]: E1203 07:12:40.429693 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105df626-6fb4-453a-9851-836d6d615e6a" containerName="init" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.430284 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="105df626-6fb4-453a-9851-836d6d615e6a" containerName="init" Dec 03 07:12:40 crc kubenswrapper[4946]: E1203 07:12:40.430353 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105df626-6fb4-453a-9851-836d6d615e6a" containerName="dnsmasq-dns" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.430425 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="105df626-6fb4-453a-9851-836d6d615e6a" containerName="dnsmasq-dns" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.430642 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="105df626-6fb4-453a-9851-836d6d615e6a" containerName="dnsmasq-dns" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.432071 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.438399 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6786b69d4-clggf"] Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.441393 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.442023 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.469382 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.513817 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/761f2199-bf20-487a-a18d-172c65a4b515-logs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.513909 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data-custom\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.513940 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmwrc\" (UniqueName: \"kubernetes.io/projected/761f2199-bf20-487a-a18d-172c65a4b515-kube-api-access-xmwrc\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.514017 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-public-tls-certs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.514079 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-internal-tls-certs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.515566 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-combined-ca-bundle\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.515603 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.617291 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.617672 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/761f2199-bf20-487a-a18d-172c65a4b515-logs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.617699 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data-custom\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.617731 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmwrc\" (UniqueName: \"kubernetes.io/projected/761f2199-bf20-487a-a18d-172c65a4b515-kube-api-access-xmwrc\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.617835 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-public-tls-certs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.617909 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-internal-tls-certs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.617936 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-combined-ca-bundle\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.618149 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/761f2199-bf20-487a-a18d-172c65a4b515-logs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.623572 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-internal-tls-certs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.624697 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-combined-ca-bundle\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.629967 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-public-tls-certs\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.630887 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data-custom\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.632224 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.634457 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmwrc\" (UniqueName: \"kubernetes.io/projected/761f2199-bf20-487a-a18d-172c65a4b515-kube-api-access-xmwrc\") pod \"barbican-api-6786b69d4-clggf\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.720634 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" event={"ID":"64332a7d-18bc-45c6-aad3-86c19e2979c4","Type":"ContainerStarted","Data":"4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362"} Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.720790 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.724450 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69454fc688-t54w7" event={"ID":"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541","Type":"ContainerStarted","Data":"ac9c0edaa246fce85ea02c792d4ac1c07afd3d940a4142fe3a468aaa6e107e16"} Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.725128 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.725164 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69454fc688-t54w7" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.732608 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76b8c8854b-nphm5" event={"ID":"0cf777b2-9dab-49de-9f2c-a119853437b3","Type":"ContainerStarted","Data":"f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce"} Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.733489 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.741219 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" podStartSLOduration=4.74119825 podStartE2EDuration="4.74119825s" podCreationTimestamp="2025-12-03 07:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:40.736066243 +0000 UTC m=+1353.532756352" watchObservedRunningTime="2025-12-03 07:12:40.74119825 +0000 UTC m=+1353.537888359" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.763527 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.764567 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-69454fc688-t54w7" podStartSLOduration=4.764549964 podStartE2EDuration="4.764549964s" podCreationTimestamp="2025-12-03 07:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:40.756519659 +0000 UTC m=+1353.553209758" watchObservedRunningTime="2025-12-03 07:12:40.764549964 +0000 UTC m=+1353.561240063" Dec 03 07:12:40 crc kubenswrapper[4946]: I1203 07:12:40.784224 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-76b8c8854b-nphm5" podStartSLOduration=4.784208029 podStartE2EDuration="4.784208029s" podCreationTimestamp="2025-12-03 07:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:40.775605839 +0000 UTC m=+1353.572295968" watchObservedRunningTime="2025-12-03 07:12:40.784208029 +0000 UTC m=+1353.580898138" Dec 03 07:12:41 crc kubenswrapper[4946]: I1203 07:12:41.764346 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-9twd7" event={"ID":"8e368226-3cfe-42a1-b2a3-52a53d227807","Type":"ContainerStarted","Data":"7d1e3a0daf343eed99376edb4c8bfb01cc40b826d5dc69f831d706bc8864bb79"} Dec 03 07:12:41 crc kubenswrapper[4946]: I1203 07:12:41.764676 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:41 crc kubenswrapper[4946]: I1203 07:12:41.786952 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-9twd7" podStartSLOduration=3.760497657 podStartE2EDuration="37.786933393s" podCreationTimestamp="2025-12-03 07:12:04 +0000 UTC" firstStartedPulling="2025-12-03 07:12:05.635845214 +0000 UTC m=+1318.432535323" lastFinishedPulling="2025-12-03 07:12:39.66228095 +0000 UTC m=+1352.458971059" observedRunningTime="2025-12-03 07:12:41.778879198 +0000 UTC m=+1354.575569317" watchObservedRunningTime="2025-12-03 07:12:41.786933393 +0000 UTC m=+1354.583623502" Dec 03 07:12:41 crc kubenswrapper[4946]: I1203 07:12:41.930128 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6786b69d4-clggf"] Dec 03 07:12:41 crc kubenswrapper[4946]: W1203 07:12:41.936342 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod761f2199_bf20_487a_a18d_172c65a4b515.slice/crio-bec07d9cc9516495d2dc8ed1b4d9325ba551821764e91c5734a0a55d91fa7ed5 WatchSource:0}: Error finding container bec07d9cc9516495d2dc8ed1b4d9325ba551821764e91c5734a0a55d91fa7ed5: Status 404 returned error can't find the container with id bec07d9cc9516495d2dc8ed1b4d9325ba551821764e91c5734a0a55d91fa7ed5 Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.776059 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-559fcd84cc-lhklt" event={"ID":"26ac177b-b889-43e7-888e-8ed02fb3ac72","Type":"ContainerStarted","Data":"ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f"} Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.776502 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-559fcd84cc-lhklt" event={"ID":"26ac177b-b889-43e7-888e-8ed02fb3ac72","Type":"ContainerStarted","Data":"27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060"} Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.778052 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6786b69d4-clggf" event={"ID":"761f2199-bf20-487a-a18d-172c65a4b515","Type":"ContainerStarted","Data":"364174d07480a501e011d47109a2fdb307a389eee25a760b44b2337057f852ec"} Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.778092 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6786b69d4-clggf" event={"ID":"761f2199-bf20-487a-a18d-172c65a4b515","Type":"ContainerStarted","Data":"29a3b6fe73a28a60ca8a3174c922f9babad002d47efe3d41eecf7fee894fc739"} Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.778110 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6786b69d4-clggf" event={"ID":"761f2199-bf20-487a-a18d-172c65a4b515","Type":"ContainerStarted","Data":"bec07d9cc9516495d2dc8ed1b4d9325ba551821764e91c5734a0a55d91fa7ed5"} Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.778132 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.778164 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.781244 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" event={"ID":"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e","Type":"ContainerStarted","Data":"4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f"} Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.781280 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" event={"ID":"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e","Type":"ContainerStarted","Data":"444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510"} Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.801349 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-559fcd84cc-lhklt" podStartSLOduration=3.666810839 podStartE2EDuration="6.80133389s" podCreationTimestamp="2025-12-03 07:12:36 +0000 UTC" firstStartedPulling="2025-12-03 07:12:38.394377841 +0000 UTC m=+1351.191067950" lastFinishedPulling="2025-12-03 07:12:41.528900882 +0000 UTC m=+1354.325591001" observedRunningTime="2025-12-03 07:12:42.797096837 +0000 UTC m=+1355.593786946" watchObservedRunningTime="2025-12-03 07:12:42.80133389 +0000 UTC m=+1355.598023999" Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.822960 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6786b69d4-clggf" podStartSLOduration=2.822940017 podStartE2EDuration="2.822940017s" podCreationTimestamp="2025-12-03 07:12:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:42.813433473 +0000 UTC m=+1355.610123583" watchObservedRunningTime="2025-12-03 07:12:42.822940017 +0000 UTC m=+1355.619630126" Dec 03 07:12:42 crc kubenswrapper[4946]: I1203 07:12:42.836906 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" podStartSLOduration=3.6850716180000003 podStartE2EDuration="6.83688738s" podCreationTimestamp="2025-12-03 07:12:36 +0000 UTC" firstStartedPulling="2025-12-03 07:12:38.378267541 +0000 UTC m=+1351.174957640" lastFinishedPulling="2025-12-03 07:12:41.530083293 +0000 UTC m=+1354.326773402" observedRunningTime="2025-12-03 07:12:42.832231946 +0000 UTC m=+1355.628922055" watchObservedRunningTime="2025-12-03 07:12:42.83688738 +0000 UTC m=+1355.633577489" Dec 03 07:12:45 crc kubenswrapper[4946]: I1203 07:12:45.806421 4946 generic.go:334] "Generic (PLEG): container finished" podID="8e368226-3cfe-42a1-b2a3-52a53d227807" containerID="7d1e3a0daf343eed99376edb4c8bfb01cc40b826d5dc69f831d706bc8864bb79" exitCode=0 Dec 03 07:12:45 crc kubenswrapper[4946]: I1203 07:12:45.806957 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-9twd7" event={"ID":"8e368226-3cfe-42a1-b2a3-52a53d227807","Type":"ContainerDied","Data":"7d1e3a0daf343eed99376edb4c8bfb01cc40b826d5dc69f831d706bc8864bb79"} Dec 03 07:12:47 crc kubenswrapper[4946]: E1203 07:12:47.073594 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.252312 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.363836 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-combined-ca-bundle\") pod \"8e368226-3cfe-42a1-b2a3-52a53d227807\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.364063 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-db-sync-config-data\") pod \"8e368226-3cfe-42a1-b2a3-52a53d227807\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.364104 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e368226-3cfe-42a1-b2a3-52a53d227807-etc-machine-id\") pod \"8e368226-3cfe-42a1-b2a3-52a53d227807\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.364144 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-scripts\") pod \"8e368226-3cfe-42a1-b2a3-52a53d227807\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.364218 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x8g8\" (UniqueName: \"kubernetes.io/projected/8e368226-3cfe-42a1-b2a3-52a53d227807-kube-api-access-8x8g8\") pod \"8e368226-3cfe-42a1-b2a3-52a53d227807\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.364231 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e368226-3cfe-42a1-b2a3-52a53d227807-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8e368226-3cfe-42a1-b2a3-52a53d227807" (UID: "8e368226-3cfe-42a1-b2a3-52a53d227807"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.364255 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-config-data\") pod \"8e368226-3cfe-42a1-b2a3-52a53d227807\" (UID: \"8e368226-3cfe-42a1-b2a3-52a53d227807\") " Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.365323 4946 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e368226-3cfe-42a1-b2a3-52a53d227807-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.369443 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8e368226-3cfe-42a1-b2a3-52a53d227807" (UID: "8e368226-3cfe-42a1-b2a3-52a53d227807"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.369844 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-scripts" (OuterVolumeSpecName: "scripts") pod "8e368226-3cfe-42a1-b2a3-52a53d227807" (UID: "8e368226-3cfe-42a1-b2a3-52a53d227807"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.372666 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e368226-3cfe-42a1-b2a3-52a53d227807-kube-api-access-8x8g8" (OuterVolumeSpecName: "kube-api-access-8x8g8") pod "8e368226-3cfe-42a1-b2a3-52a53d227807" (UID: "8e368226-3cfe-42a1-b2a3-52a53d227807"). InnerVolumeSpecName "kube-api-access-8x8g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.415201 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e368226-3cfe-42a1-b2a3-52a53d227807" (UID: "8e368226-3cfe-42a1-b2a3-52a53d227807"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.430036 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-config-data" (OuterVolumeSpecName: "config-data") pod "8e368226-3cfe-42a1-b2a3-52a53d227807" (UID: "8e368226-3cfe-42a1-b2a3-52a53d227807"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.466624 4946 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.466661 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.466675 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x8g8\" (UniqueName: \"kubernetes.io/projected/8e368226-3cfe-42a1-b2a3-52a53d227807-kube-api-access-8x8g8\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.466687 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.466699 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e368226-3cfe-42a1-b2a3-52a53d227807-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.482912 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.570220 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd5694b85-wl6d8"] Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.570873 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" podUID="1c8deacd-463c-496a-8521-de378c09ba06" containerName="dnsmasq-dns" containerID="cri-o://41f29978308af79c28f1ea282a318bc72fb22a3e70de9e072d4cf64063188c1e" gracePeriod=10 Dec 03 07:12:47 crc kubenswrapper[4946]: E1203 07:12:47.799582 4946 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c8deacd_463c_496a_8521_de378c09ba06.slice/crio-41f29978308af79c28f1ea282a318bc72fb22a3e70de9e072d4cf64063188c1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c8deacd_463c_496a_8521_de378c09ba06.slice/crio-conmon-41f29978308af79c28f1ea282a318bc72fb22a3e70de9e072d4cf64063188c1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e368226_3cfe_42a1_b2a3_52a53d227807.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e368226_3cfe_42a1_b2a3_52a53d227807.slice/crio-2e7baa520f177e39464d54a6ddd50ea9677b11f3b55b8bcc888800e2769db65e\": RecentStats: unable to find data in memory cache]" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.836352 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-9twd7" event={"ID":"8e368226-3cfe-42a1-b2a3-52a53d227807","Type":"ContainerDied","Data":"2e7baa520f177e39464d54a6ddd50ea9677b11f3b55b8bcc888800e2769db65e"} Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.836389 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e7baa520f177e39464d54a6ddd50ea9677b11f3b55b8bcc888800e2769db65e" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.836439 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-9twd7" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.840828 4946 generic.go:334] "Generic (PLEG): container finished" podID="1c8deacd-463c-496a-8521-de378c09ba06" containerID="41f29978308af79c28f1ea282a318bc72fb22a3e70de9e072d4cf64063188c1e" exitCode=0 Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.840868 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" event={"ID":"1c8deacd-463c-496a-8521-de378c09ba06","Type":"ContainerDied","Data":"41f29978308af79c28f1ea282a318bc72fb22a3e70de9e072d4cf64063188c1e"} Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.842829 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e32c0a4-449e-4f0f-9ab4-d020d313a99b","Type":"ContainerStarted","Data":"bba05f781e5eff3ec16db1987b6e068959cbbf918794ac4918886b81ef9ab623"} Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.842964 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="ceilometer-notification-agent" containerID="cri-o://61ea719a4977bd323e741afe2c8d71bdfb58887990345513124795cd4a6f53a9" gracePeriod=30 Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.843142 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.843187 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="proxy-httpd" containerID="cri-o://bba05f781e5eff3ec16db1987b6e068959cbbf918794ac4918886b81ef9ab623" gracePeriod=30 Dec 03 07:12:47 crc kubenswrapper[4946]: I1203 07:12:47.843224 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="sg-core" containerID="cri-o://c55e987e9d86be63d4718b1bc16fe9a74d06571d798460e0883408639260cec0" gracePeriod=30 Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.134601 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:12:48 crc kubenswrapper[4946]: E1203 07:12:48.136253 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e368226-3cfe-42a1-b2a3-52a53d227807" containerName="cinder-db-sync" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.136323 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e368226-3cfe-42a1-b2a3-52a53d227807" containerName="cinder-db-sync" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.136601 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e368226-3cfe-42a1-b2a3-52a53d227807" containerName="cinder-db-sync" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.137714 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.137862 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.141863 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.141875 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.141912 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-c2hp5" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.141943 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.178662 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.183550 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-sb\") pod \"1c8deacd-463c-496a-8521-de378c09ba06\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.183909 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-config\") pod \"1c8deacd-463c-496a-8521-de378c09ba06\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.184018 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-nb\") pod \"1c8deacd-463c-496a-8521-de378c09ba06\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.184202 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9t2d\" (UniqueName: \"kubernetes.io/projected/1c8deacd-463c-496a-8521-de378c09ba06-kube-api-access-d9t2d\") pod \"1c8deacd-463c-496a-8521-de378c09ba06\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.184340 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-svc\") pod \"1c8deacd-463c-496a-8521-de378c09ba06\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.184452 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-swift-storage-0\") pod \"1c8deacd-463c-496a-8521-de378c09ba06\" (UID: \"1c8deacd-463c-496a-8521-de378c09ba06\") " Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.184795 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.184916 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.185037 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.185119 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5svjk\" (UniqueName: \"kubernetes.io/projected/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-kube-api-access-5svjk\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.185441 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.185578 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.207850 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c8deacd-463c-496a-8521-de378c09ba06-kube-api-access-d9t2d" (OuterVolumeSpecName: "kube-api-access-d9t2d") pod "1c8deacd-463c-496a-8521-de378c09ba06" (UID: "1c8deacd-463c-496a-8521-de378c09ba06"). InnerVolumeSpecName "kube-api-access-d9t2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.217250 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-894d58c65-lxppn"] Dec 03 07:12:48 crc kubenswrapper[4946]: E1203 07:12:48.217730 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c8deacd-463c-496a-8521-de378c09ba06" containerName="dnsmasq-dns" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.217771 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c8deacd-463c-496a-8521-de378c09ba06" containerName="dnsmasq-dns" Dec 03 07:12:48 crc kubenswrapper[4946]: E1203 07:12:48.217802 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c8deacd-463c-496a-8521-de378c09ba06" containerName="init" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.217810 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c8deacd-463c-496a-8521-de378c09ba06" containerName="init" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.218031 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c8deacd-463c-496a-8521-de378c09ba06" containerName="dnsmasq-dns" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.228266 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.248708 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-894d58c65-lxppn"] Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289253 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289300 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5svjk\" (UniqueName: \"kubernetes.io/projected/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-kube-api-access-5svjk\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289344 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-sb\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289408 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-swift-storage-0\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289458 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dswlz\" (UniqueName: \"kubernetes.io/projected/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-kube-api-access-dswlz\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289637 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-config\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289715 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-svc\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289813 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.289914 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.290019 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.290058 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-nb\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.290091 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.290289 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9t2d\" (UniqueName: \"kubernetes.io/projected/1c8deacd-463c-496a-8521-de378c09ba06-kube-api-access-d9t2d\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.290415 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.317833 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5svjk\" (UniqueName: \"kubernetes.io/projected/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-kube-api-access-5svjk\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.327572 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.334725 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.354523 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.367619 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.380489 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1c8deacd-463c-496a-8521-de378c09ba06" (UID: "1c8deacd-463c-496a-8521-de378c09ba06"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.381844 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.392919 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-sb\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.393156 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-swift-storage-0\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.393244 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dswlz\" (UniqueName: \"kubernetes.io/projected/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-kube-api-access-dswlz\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.393357 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-config\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.393540 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-svc\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.393677 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-nb\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.393834 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.394581 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-nb\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.395390 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-config\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.395470 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-sb\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.396197 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-svc\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.396324 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-swift-storage-0\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.397469 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.397573 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.399759 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-config" (OuterVolumeSpecName: "config") pod "1c8deacd-463c-496a-8521-de378c09ba06" (UID: "1c8deacd-463c-496a-8521-de378c09ba06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.400270 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.410209 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1c8deacd-463c-496a-8521-de378c09ba06" (UID: "1c8deacd-463c-496a-8521-de378c09ba06"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.410509 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1c8deacd-463c-496a-8521-de378c09ba06" (UID: "1c8deacd-463c-496a-8521-de378c09ba06"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.416670 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dswlz\" (UniqueName: \"kubernetes.io/projected/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-kube-api-access-dswlz\") pod \"dnsmasq-dns-894d58c65-lxppn\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.443682 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1c8deacd-463c-496a-8521-de378c09ba06" (UID: "1c8deacd-463c-496a-8521-de378c09ba06"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.454754 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.495531 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-scripts\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498143 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498182 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498570 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498626 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h8pn\" (UniqueName: \"kubernetes.io/projected/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-kube-api-access-2h8pn\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498681 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-logs\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498721 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498905 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498927 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498942 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.498953 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c8deacd-463c-496a-8521-de378c09ba06-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.589215 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.600701 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.600801 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h8pn\" (UniqueName: \"kubernetes.io/projected/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-kube-api-access-2h8pn\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.600825 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-logs\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.600846 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.600879 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-scripts\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.600927 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.600942 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.603888 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.604246 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-logs\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.618333 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-scripts\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.618800 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.622695 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.624704 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h8pn\" (UniqueName: \"kubernetes.io/projected/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-kube-api-access-2h8pn\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.624809 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data\") pod \"cinder-api-0\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.721279 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.868106 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.868485 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5694b85-wl6d8" event={"ID":"1c8deacd-463c-496a-8521-de378c09ba06","Type":"ContainerDied","Data":"41c4df89e8715ff7049b467c3fb5d382eb9f1386265a7fb68b4ddea1388f73da"} Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.868533 4946 scope.go:117] "RemoveContainer" containerID="41f29978308af79c28f1ea282a318bc72fb22a3e70de9e072d4cf64063188c1e" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.876590 4946 generic.go:334] "Generic (PLEG): container finished" podID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerID="bba05f781e5eff3ec16db1987b6e068959cbbf918794ac4918886b81ef9ab623" exitCode=0 Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.876616 4946 generic.go:334] "Generic (PLEG): container finished" podID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerID="c55e987e9d86be63d4718b1bc16fe9a74d06571d798460e0883408639260cec0" exitCode=2 Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.876635 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e32c0a4-449e-4f0f-9ab4-d020d313a99b","Type":"ContainerDied","Data":"bba05f781e5eff3ec16db1987b6e068959cbbf918794ac4918886b81ef9ab623"} Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.876658 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e32c0a4-449e-4f0f-9ab4-d020d313a99b","Type":"ContainerDied","Data":"c55e987e9d86be63d4718b1bc16fe9a74d06571d798460e0883408639260cec0"} Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.910088 4946 scope.go:117] "RemoveContainer" containerID="f8e7f32fa0ed4a638857013ca3e0d38b41eaadf0d05fcfc3fb9ad097dd5e8b11" Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.916758 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd5694b85-wl6d8"] Dec 03 07:12:48 crc kubenswrapper[4946]: I1203 07:12:48.932756 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bd5694b85-wl6d8"] Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.019685 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.275274 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:49 crc kubenswrapper[4946]: W1203 07:12:49.294194 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ef1e8cc_4747_4996_8e5c_5c3f31f1b5fc.slice/crio-8cded015591d245505102a682dce67a1a4a288f981dcd3ac3800136d8ad55a59 WatchSource:0}: Error finding container 8cded015591d245505102a682dce67a1a4a288f981dcd3ac3800136d8ad55a59: Status 404 returned error can't find the container with id 8cded015591d245505102a682dce67a1a4a288f981dcd3ac3800136d8ad55a59 Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.299472 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-894d58c65-lxppn"] Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.609620 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c8deacd-463c-496a-8521-de378c09ba06" path="/var/lib/kubelet/pods/1c8deacd-463c-496a-8521-de378c09ba06/volumes" Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.636073 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.875390 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.896260 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09","Type":"ContainerStarted","Data":"91cad838c0e527f3b8d4a0562a6079467e93c8c38d87acf71a0754032f611d6f"} Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.898235 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2976bd9b-9ccd-481e-b6c3-5094d7fab70e","Type":"ContainerStarted","Data":"18cfabb679996767178dc8e1e563fab4a3c6a7d21805deaf5162c26905f733fc"} Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.902436 4946 generic.go:334] "Generic (PLEG): container finished" podID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" containerID="4f80294365bb9ae4ed7c0a9ed42dbe052d5bad38103f1a6990bc0abf4581181f" exitCode=0 Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.902486 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-894d58c65-lxppn" event={"ID":"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc","Type":"ContainerDied","Data":"4f80294365bb9ae4ed7c0a9ed42dbe052d5bad38103f1a6990bc0abf4581181f"} Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.902511 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-894d58c65-lxppn" event={"ID":"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc","Type":"ContainerStarted","Data":"8cded015591d245505102a682dce67a1a4a288f981dcd3ac3800136d8ad55a59"} Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.910957 4946 generic.go:334] "Generic (PLEG): container finished" podID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerID="61ea719a4977bd323e741afe2c8d71bdfb58887990345513124795cd4a6f53a9" exitCode=0 Dec 03 07:12:49 crc kubenswrapper[4946]: I1203 07:12:49.911676 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e32c0a4-449e-4f0f-9ab4-d020d313a99b","Type":"ContainerDied","Data":"61ea719a4977bd323e741afe2c8d71bdfb58887990345513124795cd4a6f53a9"} Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.272724 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.452422 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-config-data\") pod \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.452451 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-run-httpd\") pod \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.452507 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-combined-ca-bundle\") pod \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.452547 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm8pc\" (UniqueName: \"kubernetes.io/projected/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-kube-api-access-xm8pc\") pod \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.452602 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-scripts\") pod \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.452624 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-log-httpd\") pod \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.452661 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-sg-core-conf-yaml\") pod \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\" (UID: \"4e32c0a4-449e-4f0f-9ab4-d020d313a99b\") " Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.457267 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4e32c0a4-449e-4f0f-9ab4-d020d313a99b" (UID: "4e32c0a4-449e-4f0f-9ab4-d020d313a99b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.457428 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4e32c0a4-449e-4f0f-9ab4-d020d313a99b" (UID: "4e32c0a4-449e-4f0f-9ab4-d020d313a99b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.473902 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-kube-api-access-xm8pc" (OuterVolumeSpecName: "kube-api-access-xm8pc") pod "4e32c0a4-449e-4f0f-9ab4-d020d313a99b" (UID: "4e32c0a4-449e-4f0f-9ab4-d020d313a99b"). InnerVolumeSpecName "kube-api-access-xm8pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.475805 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-scripts" (OuterVolumeSpecName: "scripts") pod "4e32c0a4-449e-4f0f-9ab4-d020d313a99b" (UID: "4e32c0a4-449e-4f0f-9ab4-d020d313a99b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.496206 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4e32c0a4-449e-4f0f-9ab4-d020d313a99b" (UID: "4e32c0a4-449e-4f0f-9ab4-d020d313a99b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.533275 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e32c0a4-449e-4f0f-9ab4-d020d313a99b" (UID: "4e32c0a4-449e-4f0f-9ab4-d020d313a99b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.559335 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.559365 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm8pc\" (UniqueName: \"kubernetes.io/projected/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-kube-api-access-xm8pc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.559376 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.559385 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.559395 4946 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.559403 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.566936 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-config-data" (OuterVolumeSpecName: "config-data") pod "4e32c0a4-449e-4f0f-9ab4-d020d313a99b" (UID: "4e32c0a4-449e-4f0f-9ab4-d020d313a99b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.661571 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e32c0a4-449e-4f0f-9ab4-d020d313a99b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:50 crc kubenswrapper[4946]: I1203 07:12:50.957935 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09","Type":"ContainerStarted","Data":"fb006883ad5823d57304aefc470852592a0dd5a7c08986ef3a95869ef5bb0bb9"} Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.025137 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-894d58c65-lxppn" event={"ID":"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc","Type":"ContainerStarted","Data":"71e308f35c992bfb97c644a69f108037c01605c3e14403a29286ed29fc5a3878"} Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.026814 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.034708 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e32c0a4-449e-4f0f-9ab4-d020d313a99b","Type":"ContainerDied","Data":"d8d84d715a68dabefe658a5385ed7fdb8da81a3f83b33daa97de451ac1fc8803"} Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.034782 4946 scope.go:117] "RemoveContainer" containerID="bba05f781e5eff3ec16db1987b6e068959cbbf918794ac4918886b81ef9ab623" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.034981 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.116470 4946 scope.go:117] "RemoveContainer" containerID="c55e987e9d86be63d4718b1bc16fe9a74d06571d798460e0883408639260cec0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.130390 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-894d58c65-lxppn" podStartSLOduration=3.130368636 podStartE2EDuration="3.130368636s" podCreationTimestamp="2025-12-03 07:12:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:51.048302704 +0000 UTC m=+1363.844992813" watchObservedRunningTime="2025-12-03 07:12:51.130368636 +0000 UTC m=+1363.927058745" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.157454 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.182583 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.207342 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:51 crc kubenswrapper[4946]: E1203 07:12:51.207906 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="ceilometer-notification-agent" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.207926 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="ceilometer-notification-agent" Dec 03 07:12:51 crc kubenswrapper[4946]: E1203 07:12:51.207941 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="sg-core" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.207948 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="sg-core" Dec 03 07:12:51 crc kubenswrapper[4946]: E1203 07:12:51.207968 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="proxy-httpd" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.207976 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="proxy-httpd" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.208198 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="ceilometer-notification-agent" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.208219 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="sg-core" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.208243 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" containerName="proxy-httpd" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.210373 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.217749 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.221859 4946 scope.go:117] "RemoveContainer" containerID="61ea719a4977bd323e741afe2c8d71bdfb58887990345513124795cd4a6f53a9" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.222368 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.222604 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.265126 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.406114 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-run-httpd\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.406159 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-scripts\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.406269 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xsmv\" (UniqueName: \"kubernetes.io/projected/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-kube-api-access-7xsmv\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.406298 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.406319 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-log-httpd\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.406350 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.406370 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-config-data\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.509344 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xsmv\" (UniqueName: \"kubernetes.io/projected/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-kube-api-access-7xsmv\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.509398 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.509416 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-log-httpd\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.509443 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.509464 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-config-data\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.509488 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-run-httpd\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.509502 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-scripts\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.510146 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-log-httpd\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.511031 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-run-httpd\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.517455 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.519278 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.520106 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-config-data\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.527488 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-scripts\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.551713 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xsmv\" (UniqueName: \"kubernetes.io/projected/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-kube-api-access-7xsmv\") pod \"ceilometer-0\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " pod="openstack/ceilometer-0" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.612507 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e32c0a4-449e-4f0f-9ab4-d020d313a99b" path="/var/lib/kubelet/pods/4e32c0a4-449e-4f0f-9ab4-d020d313a99b/volumes" Dec 03 07:12:51 crc kubenswrapper[4946]: I1203 07:12:51.665306 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:12:52 crc kubenswrapper[4946]: I1203 07:12:52.078853 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09","Type":"ContainerStarted","Data":"32101afe6117437912e1b95336cb43f6f1701f6f1dedd7bafe391663bf42e0ff"} Dec 03 07:12:52 crc kubenswrapper[4946]: I1203 07:12:52.079266 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 07:12:52 crc kubenswrapper[4946]: I1203 07:12:52.079118 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerName="cinder-api" containerID="cri-o://32101afe6117437912e1b95336cb43f6f1701f6f1dedd7bafe391663bf42e0ff" gracePeriod=30 Dec 03 07:12:52 crc kubenswrapper[4946]: I1203 07:12:52.078922 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerName="cinder-api-log" containerID="cri-o://fb006883ad5823d57304aefc470852592a0dd5a7c08986ef3a95869ef5bb0bb9" gracePeriod=30 Dec 03 07:12:52 crc kubenswrapper[4946]: I1203 07:12:52.094523 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2976bd9b-9ccd-481e-b6c3-5094d7fab70e","Type":"ContainerStarted","Data":"950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e"} Dec 03 07:12:52 crc kubenswrapper[4946]: I1203 07:12:52.107067 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.107051006 podStartE2EDuration="4.107051006s" podCreationTimestamp="2025-12-03 07:12:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:52.102196546 +0000 UTC m=+1364.898886655" watchObservedRunningTime="2025-12-03 07:12:52.107051006 +0000 UTC m=+1364.903741115" Dec 03 07:12:52 crc kubenswrapper[4946]: I1203 07:12:52.280318 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:12:52 crc kubenswrapper[4946]: I1203 07:12:52.523071 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.039516 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.039901 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.039942 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.040990 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a645c07ea34ddd66718ac1597141e393ef581a3ac2975f8c8d11ca9aa11d50ab"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.041036 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://a645c07ea34ddd66718ac1597141e393ef581a3ac2975f8c8d11ca9aa11d50ab" gracePeriod=600 Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.096794 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.177358 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-76b8c8854b-nphm5"] Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.177596 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-76b8c8854b-nphm5" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api-log" containerID="cri-o://a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f" gracePeriod=30 Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.179783 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-76b8c8854b-nphm5" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api" containerID="cri-o://f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce" gracePeriod=30 Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.181058 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerStarted","Data":"2ad123160118303cd4963f635c57fdf7b3362479e8906a77b0f999c31cdb79aa"} Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.181089 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerStarted","Data":"19c0ce2d9462a3e781cd817883f21446ee19936f252c3f4d80f3e1df292633c3"} Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.195111 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-76b8c8854b-nphm5" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": EOF" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.201468 4946 generic.go:334] "Generic (PLEG): container finished" podID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerID="32101afe6117437912e1b95336cb43f6f1701f6f1dedd7bafe391663bf42e0ff" exitCode=0 Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.201505 4946 generic.go:334] "Generic (PLEG): container finished" podID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerID="fb006883ad5823d57304aefc470852592a0dd5a7c08986ef3a95869ef5bb0bb9" exitCode=143 Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.201554 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09","Type":"ContainerDied","Data":"32101afe6117437912e1b95336cb43f6f1701f6f1dedd7bafe391663bf42e0ff"} Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.201583 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09","Type":"ContainerDied","Data":"fb006883ad5823d57304aefc470852592a0dd5a7c08986ef3a95869ef5bb0bb9"} Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.205833 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2976bd9b-9ccd-481e-b6c3-5094d7fab70e","Type":"ContainerStarted","Data":"147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750"} Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.244705 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.177105107 podStartE2EDuration="5.244687894s" podCreationTimestamp="2025-12-03 07:12:48 +0000 UTC" firstStartedPulling="2025-12-03 07:12:49.021959186 +0000 UTC m=+1361.818649295" lastFinishedPulling="2025-12-03 07:12:50.089541973 +0000 UTC m=+1362.886232082" observedRunningTime="2025-12-03 07:12:53.236728321 +0000 UTC m=+1366.033418430" watchObservedRunningTime="2025-12-03 07:12:53.244687894 +0000 UTC m=+1366.041378003" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.335358 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.455126 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.474816 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h8pn\" (UniqueName: \"kubernetes.io/projected/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-kube-api-access-2h8pn\") pod \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.474871 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-etc-machine-id\") pod \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.474924 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data-custom\") pod \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.474995 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-logs\") pod \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.475048 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data\") pod \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.475076 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-scripts\") pod \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.475149 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-combined-ca-bundle\") pod \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\" (UID: \"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09\") " Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.476056 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-logs" (OuterVolumeSpecName: "logs") pod "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" (UID: "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.476198 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" (UID: "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.483925 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-scripts" (OuterVolumeSpecName: "scripts") pod "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" (UID: "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.483958 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" (UID: "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.484063 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-kube-api-access-2h8pn" (OuterVolumeSpecName: "kube-api-access-2h8pn") pod "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" (UID: "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09"). InnerVolumeSpecName "kube-api-access-2h8pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.502112 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" (UID: "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.529349 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data" (OuterVolumeSpecName: "config-data") pod "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" (UID: "d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.577616 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.577641 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.577651 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.577660 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.577672 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h8pn\" (UniqueName: \"kubernetes.io/projected/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-kube-api-access-2h8pn\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.577681 4946 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:53 crc kubenswrapper[4946]: I1203 07:12:53.577690 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.215183 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.215363 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09","Type":"ContainerDied","Data":"91cad838c0e527f3b8d4a0562a6079467e93c8c38d87acf71a0754032f611d6f"} Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.215581 4946 scope.go:117] "RemoveContainer" containerID="32101afe6117437912e1b95336cb43f6f1701f6f1dedd7bafe391663bf42e0ff" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.217674 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="a645c07ea34ddd66718ac1597141e393ef581a3ac2975f8c8d11ca9aa11d50ab" exitCode=0 Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.217755 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"a645c07ea34ddd66718ac1597141e393ef581a3ac2975f8c8d11ca9aa11d50ab"} Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.220352 4946 generic.go:334] "Generic (PLEG): container finished" podID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerID="a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f" exitCode=143 Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.221164 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76b8c8854b-nphm5" event={"ID":"0cf777b2-9dab-49de-9f2c-a119853437b3","Type":"ContainerDied","Data":"a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f"} Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.238714 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.248092 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.254229 4946 scope.go:117] "RemoveContainer" containerID="fb006883ad5823d57304aefc470852592a0dd5a7c08986ef3a95869ef5bb0bb9" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.263650 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:54 crc kubenswrapper[4946]: E1203 07:12:54.263998 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerName="cinder-api-log" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.264014 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerName="cinder-api-log" Dec 03 07:12:54 crc kubenswrapper[4946]: E1203 07:12:54.264037 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerName="cinder-api" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.264045 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerName="cinder-api" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.264219 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerName="cinder-api-log" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.264248 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" containerName="cinder-api" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.265091 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.276720 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.277800 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.279680 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.281961 4946 scope.go:117] "RemoveContainer" containerID="b5f3af879642bc444185e03b298fe438ec767c4e61a69937761fbbcf8110a3b1" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.288374 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399284 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399343 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399368 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4xmb\" (UniqueName: \"kubernetes.io/projected/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-kube-api-access-h4xmb\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399398 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-scripts\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399432 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399553 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399586 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399609 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-logs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.399632 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501445 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501523 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501563 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501587 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-logs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501605 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501657 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501678 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501694 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4xmb\" (UniqueName: \"kubernetes.io/projected/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-kube-api-access-h4xmb\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.501717 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-scripts\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.502542 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-logs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.502613 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.507367 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.507985 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.508565 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.510258 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.511677 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.513076 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-scripts\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.521670 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4xmb\" (UniqueName: \"kubernetes.io/projected/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-kube-api-access-h4xmb\") pod \"cinder-api-0\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " pod="openstack/cinder-api-0" Dec 03 07:12:54 crc kubenswrapper[4946]: I1203 07:12:54.642573 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 07:12:55 crc kubenswrapper[4946]: I1203 07:12:55.095414 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:12:55 crc kubenswrapper[4946]: W1203 07:12:55.103909 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a3521a3_0f30_4afe_b5f8_efeb6f838e72.slice/crio-3484bdb32a49685d70086dbee64b43f466364091478c44939ca438f3cbe8d541 WatchSource:0}: Error finding container 3484bdb32a49685d70086dbee64b43f466364091478c44939ca438f3cbe8d541: Status 404 returned error can't find the container with id 3484bdb32a49685d70086dbee64b43f466364091478c44939ca438f3cbe8d541 Dec 03 07:12:55 crc kubenswrapper[4946]: I1203 07:12:55.252164 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerStarted","Data":"20c5e7246f81600dfe918cfe6ae5bd0024fc7c85452e2cf955d1de4cce76c3d8"} Dec 03 07:12:55 crc kubenswrapper[4946]: I1203 07:12:55.252211 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerStarted","Data":"9f9e599fc260887c2f3cdf78174f2cfd8105c2b2acb754d75ea473bfa9769f75"} Dec 03 07:12:55 crc kubenswrapper[4946]: I1203 07:12:55.253652 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a3521a3-0f30-4afe-b5f8-efeb6f838e72","Type":"ContainerStarted","Data":"3484bdb32a49685d70086dbee64b43f466364091478c44939ca438f3cbe8d541"} Dec 03 07:12:55 crc kubenswrapper[4946]: I1203 07:12:55.260003 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113"} Dec 03 07:12:55 crc kubenswrapper[4946]: I1203 07:12:55.638820 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09" path="/var/lib/kubelet/pods/d4c2cce3-d10f-4b2e-8ca7-f86f038f8e09/volumes" Dec 03 07:12:56 crc kubenswrapper[4946]: I1203 07:12:56.277067 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a3521a3-0f30-4afe-b5f8-efeb6f838e72","Type":"ContainerStarted","Data":"0c1e1e5afa702c1d51ea8d3cbdd531ef58625215a8edb86478cd68e3ae7a611f"} Dec 03 07:12:57 crc kubenswrapper[4946]: I1203 07:12:57.287362 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a3521a3-0f30-4afe-b5f8-efeb6f838e72","Type":"ContainerStarted","Data":"6d8145ae8f7303d9ebe9697e32c04cdfb3a46dd8ce1dc51d2f81b3c03b54e2c8"} Dec 03 07:12:57 crc kubenswrapper[4946]: I1203 07:12:57.292816 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 07:12:57 crc kubenswrapper[4946]: I1203 07:12:57.295380 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerStarted","Data":"b8a4e58ce6d08454436a5eca52e0c2ee12e584df13a439c59b479ea93d5a27f4"} Dec 03 07:12:57 crc kubenswrapper[4946]: I1203 07:12:57.295584 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 07:12:57 crc kubenswrapper[4946]: I1203 07:12:57.346766 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.346725528 podStartE2EDuration="3.346725528s" podCreationTimestamp="2025-12-03 07:12:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:12:57.325015478 +0000 UTC m=+1370.121705607" watchObservedRunningTime="2025-12-03 07:12:57.346725528 +0000 UTC m=+1370.143415637" Dec 03 07:12:57 crc kubenswrapper[4946]: I1203 07:12:57.351824 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.362359658 podStartE2EDuration="6.351795373s" podCreationTimestamp="2025-12-03 07:12:51 +0000 UTC" firstStartedPulling="2025-12-03 07:12:52.274459458 +0000 UTC m=+1365.071149567" lastFinishedPulling="2025-12-03 07:12:56.263895173 +0000 UTC m=+1369.060585282" observedRunningTime="2025-12-03 07:12:57.344710564 +0000 UTC m=+1370.141400683" watchObservedRunningTime="2025-12-03 07:12:57.351795373 +0000 UTC m=+1370.148485482" Dec 03 07:12:57 crc kubenswrapper[4946]: I1203 07:12:57.626945 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76b8c8854b-nphm5" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": read tcp 10.217.0.2:38844->10.217.0.155:9311: read: connection reset by peer" Dec 03 07:12:57 crc kubenswrapper[4946]: I1203 07:12:57.626962 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76b8c8854b-nphm5" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": read tcp 10.217.0.2:38856->10.217.0.155:9311: read: connection reset by peer" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.130379 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.292394 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data-custom\") pod \"0cf777b2-9dab-49de-9f2c-a119853437b3\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.292495 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxr4t\" (UniqueName: \"kubernetes.io/projected/0cf777b2-9dab-49de-9f2c-a119853437b3-kube-api-access-lxr4t\") pod \"0cf777b2-9dab-49de-9f2c-a119853437b3\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.292529 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data\") pod \"0cf777b2-9dab-49de-9f2c-a119853437b3\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.292593 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-combined-ca-bundle\") pod \"0cf777b2-9dab-49de-9f2c-a119853437b3\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.292633 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cf777b2-9dab-49de-9f2c-a119853437b3-logs\") pod \"0cf777b2-9dab-49de-9f2c-a119853437b3\" (UID: \"0cf777b2-9dab-49de-9f2c-a119853437b3\") " Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.294026 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cf777b2-9dab-49de-9f2c-a119853437b3-logs" (OuterVolumeSpecName: "logs") pod "0cf777b2-9dab-49de-9f2c-a119853437b3" (UID: "0cf777b2-9dab-49de-9f2c-a119853437b3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.301936 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cf777b2-9dab-49de-9f2c-a119853437b3-kube-api-access-lxr4t" (OuterVolumeSpecName: "kube-api-access-lxr4t") pod "0cf777b2-9dab-49de-9f2c-a119853437b3" (UID: "0cf777b2-9dab-49de-9f2c-a119853437b3"). InnerVolumeSpecName "kube-api-access-lxr4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.310410 4946 generic.go:334] "Generic (PLEG): container finished" podID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerID="f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce" exitCode=0 Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.310575 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76b8c8854b-nphm5" event={"ID":"0cf777b2-9dab-49de-9f2c-a119853437b3","Type":"ContainerDied","Data":"f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce"} Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.310763 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76b8c8854b-nphm5" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.310909 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76b8c8854b-nphm5" event={"ID":"0cf777b2-9dab-49de-9f2c-a119853437b3","Type":"ContainerDied","Data":"473d31036a45bddce221695b524d1dc6fe3ed1343117b99b649f7a6754c81e8d"} Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.311243 4946 scope.go:117] "RemoveContainer" containerID="f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.318908 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0cf777b2-9dab-49de-9f2c-a119853437b3" (UID: "0cf777b2-9dab-49de-9f2c-a119853437b3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.342391 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cf777b2-9dab-49de-9f2c-a119853437b3" (UID: "0cf777b2-9dab-49de-9f2c-a119853437b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.367081 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data" (OuterVolumeSpecName: "config-data") pod "0cf777b2-9dab-49de-9f2c-a119853437b3" (UID: "0cf777b2-9dab-49de-9f2c-a119853437b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.395370 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.395417 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cf777b2-9dab-49de-9f2c-a119853437b3-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.395430 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.395442 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxr4t\" (UniqueName: \"kubernetes.io/projected/0cf777b2-9dab-49de-9f2c-a119853437b3-kube-api-access-lxr4t\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.395456 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf777b2-9dab-49de-9f2c-a119853437b3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.445644 4946 scope.go:117] "RemoveContainer" containerID="a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.471348 4946 scope.go:117] "RemoveContainer" containerID="f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce" Dec 03 07:12:58 crc kubenswrapper[4946]: E1203 07:12:58.471865 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce\": container with ID starting with f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce not found: ID does not exist" containerID="f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.471944 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce"} err="failed to get container status \"f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce\": rpc error: code = NotFound desc = could not find container \"f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce\": container with ID starting with f18344f1eb1f823fcc0e313c286a8470d10eda8bbb3a67e6e865fbf1faed60ce not found: ID does not exist" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.472000 4946 scope.go:117] "RemoveContainer" containerID="a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f" Dec 03 07:12:58 crc kubenswrapper[4946]: E1203 07:12:58.472726 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f\": container with ID starting with a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f not found: ID does not exist" containerID="a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.472782 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f"} err="failed to get container status \"a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f\": rpc error: code = NotFound desc = could not find container \"a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f\": container with ID starting with a6ed84b9ddcf98f72716f6cd909c0d2f791daf4fe0b0468f2fdd3dbd1b0c3a9f not found: ID does not exist" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.592016 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.698967 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.706784 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-76b8c8854b-nphm5"] Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.743711 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f568c56f7-69bfr"] Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.744159 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" podUID="64332a7d-18bc-45c6-aad3-86c19e2979c4" containerName="dnsmasq-dns" containerID="cri-o://4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362" gracePeriod=10 Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.770409 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-76b8c8854b-nphm5"] Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.831079 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:12:58 crc kubenswrapper[4946]: I1203 07:12:58.844010 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.298037 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.327520 4946 generic.go:334] "Generic (PLEG): container finished" podID="64332a7d-18bc-45c6-aad3-86c19e2979c4" containerID="4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362" exitCode=0 Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.327590 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" event={"ID":"64332a7d-18bc-45c6-aad3-86c19e2979c4","Type":"ContainerDied","Data":"4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362"} Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.327625 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" event={"ID":"64332a7d-18bc-45c6-aad3-86c19e2979c4","Type":"ContainerDied","Data":"0bd8bbc572e17859fac80900b57065a3dcf824272f38b7df050320fd21279e40"} Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.327643 4946 scope.go:117] "RemoveContainer" containerID="4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.327641 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f568c56f7-69bfr" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.328610 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerName="cinder-scheduler" containerID="cri-o://950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e" gracePeriod=30 Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.328776 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerName="probe" containerID="cri-o://147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750" gracePeriod=30 Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.360625 4946 scope.go:117] "RemoveContainer" containerID="362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.379376 4946 scope.go:117] "RemoveContainer" containerID="4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362" Dec 03 07:12:59 crc kubenswrapper[4946]: E1203 07:12:59.379814 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362\": container with ID starting with 4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362 not found: ID does not exist" containerID="4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.379849 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362"} err="failed to get container status \"4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362\": rpc error: code = NotFound desc = could not find container \"4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362\": container with ID starting with 4ff92cd8b5f9857cc41bad54e9ba4a6331571aa3f157044a0b9bb1a51a354362 not found: ID does not exist" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.379872 4946 scope.go:117] "RemoveContainer" containerID="362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a" Dec 03 07:12:59 crc kubenswrapper[4946]: E1203 07:12:59.380240 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a\": container with ID starting with 362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a not found: ID does not exist" containerID="362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.380263 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a"} err="failed to get container status \"362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a\": rpc error: code = NotFound desc = could not find container \"362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a\": container with ID starting with 362912388fb2cb673fa8c189899a9c603a4e1a95c0b8748b0fa890c43568947a not found: ID does not exist" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.423814 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-svc\") pod \"64332a7d-18bc-45c6-aad3-86c19e2979c4\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.423901 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-nb\") pod \"64332a7d-18bc-45c6-aad3-86c19e2979c4\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.423930 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-config\") pod \"64332a7d-18bc-45c6-aad3-86c19e2979c4\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.423994 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-swift-storage-0\") pod \"64332a7d-18bc-45c6-aad3-86c19e2979c4\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.424099 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-sb\") pod \"64332a7d-18bc-45c6-aad3-86c19e2979c4\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.424129 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jxfz\" (UniqueName: \"kubernetes.io/projected/64332a7d-18bc-45c6-aad3-86c19e2979c4-kube-api-access-5jxfz\") pod \"64332a7d-18bc-45c6-aad3-86c19e2979c4\" (UID: \"64332a7d-18bc-45c6-aad3-86c19e2979c4\") " Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.431653 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64332a7d-18bc-45c6-aad3-86c19e2979c4-kube-api-access-5jxfz" (OuterVolumeSpecName: "kube-api-access-5jxfz") pod "64332a7d-18bc-45c6-aad3-86c19e2979c4" (UID: "64332a7d-18bc-45c6-aad3-86c19e2979c4"). InnerVolumeSpecName "kube-api-access-5jxfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.481931 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "64332a7d-18bc-45c6-aad3-86c19e2979c4" (UID: "64332a7d-18bc-45c6-aad3-86c19e2979c4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.492516 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "64332a7d-18bc-45c6-aad3-86c19e2979c4" (UID: "64332a7d-18bc-45c6-aad3-86c19e2979c4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.493110 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-config" (OuterVolumeSpecName: "config") pod "64332a7d-18bc-45c6-aad3-86c19e2979c4" (UID: "64332a7d-18bc-45c6-aad3-86c19e2979c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.497208 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "64332a7d-18bc-45c6-aad3-86c19e2979c4" (UID: "64332a7d-18bc-45c6-aad3-86c19e2979c4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.503910 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "64332a7d-18bc-45c6-aad3-86c19e2979c4" (UID: "64332a7d-18bc-45c6-aad3-86c19e2979c4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.528030 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.528063 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.528073 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.528085 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.528097 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/64332a7d-18bc-45c6-aad3-86c19e2979c4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.528107 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jxfz\" (UniqueName: \"kubernetes.io/projected/64332a7d-18bc-45c6-aad3-86c19e2979c4-kube-api-access-5jxfz\") on node \"crc\" DevicePath \"\"" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.603893 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" path="/var/lib/kubelet/pods/0cf777b2-9dab-49de-9f2c-a119853437b3/volumes" Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.657478 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f568c56f7-69bfr"] Dec 03 07:12:59 crc kubenswrapper[4946]: I1203 07:12:59.666612 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f568c56f7-69bfr"] Dec 03 07:13:00 crc kubenswrapper[4946]: I1203 07:13:00.339376 4946 generic.go:334] "Generic (PLEG): container finished" podID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerID="147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750" exitCode=0 Dec 03 07:13:00 crc kubenswrapper[4946]: I1203 07:13:00.339472 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2976bd9b-9ccd-481e-b6c3-5094d7fab70e","Type":"ContainerDied","Data":"147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750"} Dec 03 07:13:01 crc kubenswrapper[4946]: I1203 07:13:01.610207 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64332a7d-18bc-45c6-aad3-86c19e2979c4" path="/var/lib/kubelet/pods/64332a7d-18bc-45c6-aad3-86c19e2979c4/volumes" Dec 03 07:13:01 crc kubenswrapper[4946]: I1203 07:13:01.688858 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:13:01 crc kubenswrapper[4946]: I1203 07:13:01.767171 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6945f689f6-c4fbq"] Dec 03 07:13:01 crc kubenswrapper[4946]: I1203 07:13:01.768091 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6945f689f6-c4fbq" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerName="neutron-api" containerID="cri-o://bbf789b9f76a02d9d9c15d2b2bb403dd915d29bc873e449ba1e2b5fde5777c80" gracePeriod=30 Dec 03 07:13:01 crc kubenswrapper[4946]: I1203 07:13:01.768691 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6945f689f6-c4fbq" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerName="neutron-httpd" containerID="cri-o://7e5a02adf9b400b9fc12a49adbdc155b358023367bd475eefec73c51efc530e7" gracePeriod=30 Dec 03 07:13:02 crc kubenswrapper[4946]: I1203 07:13:02.365656 4946 generic.go:334] "Generic (PLEG): container finished" podID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerID="7e5a02adf9b400b9fc12a49adbdc155b358023367bd475eefec73c51efc530e7" exitCode=0 Dec 03 07:13:02 crc kubenswrapper[4946]: I1203 07:13:02.365771 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6945f689f6-c4fbq" event={"ID":"9c80a231-30ed-4e17-b8ba-20b628b8b9bd","Type":"ContainerDied","Data":"7e5a02adf9b400b9fc12a49adbdc155b358023367bd475eefec73c51efc530e7"} Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.296565 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.382074 4946 generic.go:334] "Generic (PLEG): container finished" podID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerID="950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e" exitCode=0 Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.382480 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2976bd9b-9ccd-481e-b6c3-5094d7fab70e","Type":"ContainerDied","Data":"950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e"} Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.382523 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2976bd9b-9ccd-481e-b6c3-5094d7fab70e","Type":"ContainerDied","Data":"18cfabb679996767178dc8e1e563fab4a3c6a7d21805deaf5162c26905f733fc"} Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.382551 4946 scope.go:117] "RemoveContainer" containerID="147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.382731 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.402319 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-etc-machine-id\") pod \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.402407 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data-custom\") pod \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.402494 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-scripts\") pod \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.402892 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2976bd9b-9ccd-481e-b6c3-5094d7fab70e" (UID: "2976bd9b-9ccd-481e-b6c3-5094d7fab70e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.402907 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-combined-ca-bundle\") pod \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.403050 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5svjk\" (UniqueName: \"kubernetes.io/projected/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-kube-api-access-5svjk\") pod \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.403130 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data\") pod \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\" (UID: \"2976bd9b-9ccd-481e-b6c3-5094d7fab70e\") " Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.403812 4946 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.416949 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-kube-api-access-5svjk" (OuterVolumeSpecName: "kube-api-access-5svjk") pod "2976bd9b-9ccd-481e-b6c3-5094d7fab70e" (UID: "2976bd9b-9ccd-481e-b6c3-5094d7fab70e"). InnerVolumeSpecName "kube-api-access-5svjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.423413 4946 scope.go:117] "RemoveContainer" containerID="950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.426215 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2976bd9b-9ccd-481e-b6c3-5094d7fab70e" (UID: "2976bd9b-9ccd-481e-b6c3-5094d7fab70e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.429458 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-scripts" (OuterVolumeSpecName: "scripts") pod "2976bd9b-9ccd-481e-b6c3-5094d7fab70e" (UID: "2976bd9b-9ccd-481e-b6c3-5094d7fab70e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.469106 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2976bd9b-9ccd-481e-b6c3-5094d7fab70e" (UID: "2976bd9b-9ccd-481e-b6c3-5094d7fab70e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.506185 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.506229 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5svjk\" (UniqueName: \"kubernetes.io/projected/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-kube-api-access-5svjk\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.506244 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.506256 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.525633 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data" (OuterVolumeSpecName: "config-data") pod "2976bd9b-9ccd-481e-b6c3-5094d7fab70e" (UID: "2976bd9b-9ccd-481e-b6c3-5094d7fab70e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.567334 4946 scope.go:117] "RemoveContainer" containerID="147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750" Dec 03 07:13:03 crc kubenswrapper[4946]: E1203 07:13:03.567941 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750\": container with ID starting with 147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750 not found: ID does not exist" containerID="147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.567987 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750"} err="failed to get container status \"147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750\": rpc error: code = NotFound desc = could not find container \"147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750\": container with ID starting with 147818856d8396401d2a1390c6c91ff6c9dc11c9157aba3c96307b24ae495750 not found: ID does not exist" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.568019 4946 scope.go:117] "RemoveContainer" containerID="950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e" Dec 03 07:13:03 crc kubenswrapper[4946]: E1203 07:13:03.568339 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e\": container with ID starting with 950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e not found: ID does not exist" containerID="950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.568376 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e"} err="failed to get container status \"950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e\": rpc error: code = NotFound desc = could not find container \"950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e\": container with ID starting with 950dbd4ff534495b9b2d7d54f9e6bf39dc528197959e3386dfa99fd9bda3ac2e not found: ID does not exist" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.608156 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2976bd9b-9ccd-481e-b6c3-5094d7fab70e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.708308 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.715293 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.742532 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:13:03 crc kubenswrapper[4946]: E1203 07:13:03.743121 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api-log" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743148 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api-log" Dec 03 07:13:03 crc kubenswrapper[4946]: E1203 07:13:03.743175 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerName="probe" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743184 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerName="probe" Dec 03 07:13:03 crc kubenswrapper[4946]: E1203 07:13:03.743199 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerName="cinder-scheduler" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743207 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerName="cinder-scheduler" Dec 03 07:13:03 crc kubenswrapper[4946]: E1203 07:13:03.743220 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64332a7d-18bc-45c6-aad3-86c19e2979c4" containerName="dnsmasq-dns" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743241 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="64332a7d-18bc-45c6-aad3-86c19e2979c4" containerName="dnsmasq-dns" Dec 03 07:13:03 crc kubenswrapper[4946]: E1203 07:13:03.743262 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743270 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api" Dec 03 07:13:03 crc kubenswrapper[4946]: E1203 07:13:03.743291 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64332a7d-18bc-45c6-aad3-86c19e2979c4" containerName="init" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743299 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="64332a7d-18bc-45c6-aad3-86c19e2979c4" containerName="init" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743509 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerName="probe" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743523 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" containerName="cinder-scheduler" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743535 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="64332a7d-18bc-45c6-aad3-86c19e2979c4" containerName="dnsmasq-dns" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743551 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.743583 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf777b2-9dab-49de-9f2c-a119853437b3" containerName="barbican-api-log" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.744814 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.747298 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.753331 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.912723 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-scripts\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.912807 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.912828 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.913127 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.913243 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60cdd367-2981-42ad-a896-23ceee7ac34e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:03 crc kubenswrapper[4946]: I1203 07:13:03.913385 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw8r7\" (UniqueName: \"kubernetes.io/projected/60cdd367-2981-42ad-a896-23ceee7ac34e-kube-api-access-bw8r7\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.015379 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-scripts\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.015452 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.015470 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.015495 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.015522 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60cdd367-2981-42ad-a896-23ceee7ac34e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.015555 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw8r7\" (UniqueName: \"kubernetes.io/projected/60cdd367-2981-42ad-a896-23ceee7ac34e-kube-api-access-bw8r7\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.015929 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60cdd367-2981-42ad-a896-23ceee7ac34e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.021257 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.021848 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.022549 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.023055 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-scripts\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.044804 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw8r7\" (UniqueName: \"kubernetes.io/projected/60cdd367-2981-42ad-a896-23ceee7ac34e-kube-api-access-bw8r7\") pod \"cinder-scheduler-0\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.095354 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 07:13:04 crc kubenswrapper[4946]: I1203 07:13:04.524432 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:13:04 crc kubenswrapper[4946]: W1203 07:13:04.537121 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60cdd367_2981_42ad_a896_23ceee7ac34e.slice/crio-42add7f84388be937cd02dd8af254dda6e580ef044a1f522802cffb491afe226 WatchSource:0}: Error finding container 42add7f84388be937cd02dd8af254dda6e580ef044a1f522802cffb491afe226: Status 404 returned error can't find the container with id 42add7f84388be937cd02dd8af254dda6e580ef044a1f522802cffb491afe226 Dec 03 07:13:05 crc kubenswrapper[4946]: I1203 07:13:05.406154 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60cdd367-2981-42ad-a896-23ceee7ac34e","Type":"ContainerStarted","Data":"fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219"} Dec 03 07:13:05 crc kubenswrapper[4946]: I1203 07:13:05.406439 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60cdd367-2981-42ad-a896-23ceee7ac34e","Type":"ContainerStarted","Data":"42add7f84388be937cd02dd8af254dda6e580ef044a1f522802cffb491afe226"} Dec 03 07:13:05 crc kubenswrapper[4946]: I1203 07:13:05.636233 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2976bd9b-9ccd-481e-b6c3-5094d7fab70e" path="/var/lib/kubelet/pods/2976bd9b-9ccd-481e-b6c3-5094d7fab70e/volumes" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.448245 4946 generic.go:334] "Generic (PLEG): container finished" podID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerID="bbf789b9f76a02d9d9c15d2b2bb403dd915d29bc873e449ba1e2b5fde5777c80" exitCode=0 Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.448419 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6945f689f6-c4fbq" event={"ID":"9c80a231-30ed-4e17-b8ba-20b628b8b9bd","Type":"ContainerDied","Data":"bbf789b9f76a02d9d9c15d2b2bb403dd915d29bc873e449ba1e2b5fde5777c80"} Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.451725 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60cdd367-2981-42ad-a896-23ceee7ac34e","Type":"ContainerStarted","Data":"ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374"} Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.516919 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.516898193 podStartE2EDuration="3.516898193s" podCreationTimestamp="2025-12-03 07:13:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:13:06.471727876 +0000 UTC m=+1379.268417985" watchObservedRunningTime="2025-12-03 07:13:06.516898193 +0000 UTC m=+1379.313588302" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.560626 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.798081 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.876975 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-ovndb-tls-certs\") pod \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.877064 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-combined-ca-bundle\") pod \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.877114 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqb7k\" (UniqueName: \"kubernetes.io/projected/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-kube-api-access-sqb7k\") pod \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.877161 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-config\") pod \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.877187 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-httpd-config\") pod \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\" (UID: \"9c80a231-30ed-4e17-b8ba-20b628b8b9bd\") " Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.884303 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-kube-api-access-sqb7k" (OuterVolumeSpecName: "kube-api-access-sqb7k") pod "9c80a231-30ed-4e17-b8ba-20b628b8b9bd" (UID: "9c80a231-30ed-4e17-b8ba-20b628b8b9bd"). InnerVolumeSpecName "kube-api-access-sqb7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.896308 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9c80a231-30ed-4e17-b8ba-20b628b8b9bd" (UID: "9c80a231-30ed-4e17-b8ba-20b628b8b9bd"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.931279 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c80a231-30ed-4e17-b8ba-20b628b8b9bd" (UID: "9c80a231-30ed-4e17-b8ba-20b628b8b9bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.953125 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-config" (OuterVolumeSpecName: "config") pod "9c80a231-30ed-4e17-b8ba-20b628b8b9bd" (UID: "9c80a231-30ed-4e17-b8ba-20b628b8b9bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.967277 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9c80a231-30ed-4e17-b8ba-20b628b8b9bd" (UID: "9c80a231-30ed-4e17-b8ba-20b628b8b9bd"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.978854 4946 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.978878 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.978887 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqb7k\" (UniqueName: \"kubernetes.io/projected/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-kube-api-access-sqb7k\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.978907 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:06 crc kubenswrapper[4946]: I1203 07:13:06.978918 4946 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9c80a231-30ed-4e17-b8ba-20b628b8b9bd-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:07 crc kubenswrapper[4946]: I1203 07:13:07.463601 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6945f689f6-c4fbq" event={"ID":"9c80a231-30ed-4e17-b8ba-20b628b8b9bd","Type":"ContainerDied","Data":"687090c41ce1ed4585057f91a2df313a47f4caef1c76cfa8f354d076c8a7deda"} Dec 03 07:13:07 crc kubenswrapper[4946]: I1203 07:13:07.463623 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6945f689f6-c4fbq" Dec 03 07:13:07 crc kubenswrapper[4946]: I1203 07:13:07.463652 4946 scope.go:117] "RemoveContainer" containerID="7e5a02adf9b400b9fc12a49adbdc155b358023367bd475eefec73c51efc530e7" Dec 03 07:13:07 crc kubenswrapper[4946]: I1203 07:13:07.495207 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6945f689f6-c4fbq"] Dec 03 07:13:07 crc kubenswrapper[4946]: I1203 07:13:07.497105 4946 scope.go:117] "RemoveContainer" containerID="bbf789b9f76a02d9d9c15d2b2bb403dd915d29bc873e449ba1e2b5fde5777c80" Dec 03 07:13:07 crc kubenswrapper[4946]: I1203 07:13:07.502887 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6945f689f6-c4fbq"] Dec 03 07:13:07 crc kubenswrapper[4946]: I1203 07:13:07.617487 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" path="/var/lib/kubelet/pods/9c80a231-30ed-4e17-b8ba-20b628b8b9bd/volumes" Dec 03 07:13:08 crc kubenswrapper[4946]: I1203 07:13:08.245312 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69454fc688-t54w7" Dec 03 07:13:08 crc kubenswrapper[4946]: I1203 07:13:08.245366 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69454fc688-t54w7" Dec 03 07:13:08 crc kubenswrapper[4946]: I1203 07:13:08.907167 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.096176 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.636641 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 07:13:09 crc kubenswrapper[4946]: E1203 07:13:09.637222 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerName="neutron-httpd" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.637245 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerName="neutron-httpd" Dec 03 07:13:09 crc kubenswrapper[4946]: E1203 07:13:09.637271 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerName="neutron-api" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.637281 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerName="neutron-api" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.637547 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerName="neutron-httpd" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.637616 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c80a231-30ed-4e17-b8ba-20b628b8b9bd" containerName="neutron-api" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.638480 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.649202 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.649524 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-h8n8l" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.649805 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.659088 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.736902 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.736992 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config-secret\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.737088 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmhdc\" (UniqueName: \"kubernetes.io/projected/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-kube-api-access-xmhdc\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.737249 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.839031 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.839096 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.839148 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config-secret\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.839229 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmhdc\" (UniqueName: \"kubernetes.io/projected/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-kube-api-access-xmhdc\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.840067 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.846785 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.858251 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config-secret\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.861678 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmhdc\" (UniqueName: \"kubernetes.io/projected/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-kube-api-access-xmhdc\") pod \"openstackclient\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.908181 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.909152 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.919296 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.951824 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.953149 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:13:09 crc kubenswrapper[4946]: I1203 07:13:09.976084 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.042753 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcb4v\" (UniqueName: \"kubernetes.io/projected/788c3bf1-3291-4324-8a4c-dc06f1027f5b-kube-api-access-hcb4v\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: E1203 07:13:10.042771 4946 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 07:13:10 crc kubenswrapper[4946]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_73f2c6a5-bed7-4e90-a70d-9fea1581bcfb_0(b54e7e4e1b65027f9d6621f2829118c6305c8b5372f886951b5f6c2e61553758): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"b54e7e4e1b65027f9d6621f2829118c6305c8b5372f886951b5f6c2e61553758" Netns:"/var/run/netns/943584ef-1536-4f18-9c56-301abaa257d1" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=b54e7e4e1b65027f9d6621f2829118c6305c8b5372f886951b5f6c2e61553758;K8S_POD_UID=73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb]: expected pod UID "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" but got "788c3bf1-3291-4324-8a4c-dc06f1027f5b" from Kube API Dec 03 07:13:10 crc kubenswrapper[4946]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 07:13:10 crc kubenswrapper[4946]: > Dec 03 07:13:10 crc kubenswrapper[4946]: E1203 07:13:10.042841 4946 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 07:13:10 crc kubenswrapper[4946]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_73f2c6a5-bed7-4e90-a70d-9fea1581bcfb_0(b54e7e4e1b65027f9d6621f2829118c6305c8b5372f886951b5f6c2e61553758): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"b54e7e4e1b65027f9d6621f2829118c6305c8b5372f886951b5f6c2e61553758" Netns:"/var/run/netns/943584ef-1536-4f18-9c56-301abaa257d1" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=b54e7e4e1b65027f9d6621f2829118c6305c8b5372f886951b5f6c2e61553758;K8S_POD_UID=73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb]: expected pod UID "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" but got "788c3bf1-3291-4324-8a4c-dc06f1027f5b" from Kube API Dec 03 07:13:10 crc kubenswrapper[4946]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 07:13:10 crc kubenswrapper[4946]: > pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.042907 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config-secret\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.042930 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.042978 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.144472 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config-secret\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.144525 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.144823 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.144955 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcb4v\" (UniqueName: \"kubernetes.io/projected/788c3bf1-3291-4324-8a4c-dc06f1027f5b-kube-api-access-hcb4v\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.145794 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.153291 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config-secret\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.153405 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.160436 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcb4v\" (UniqueName: \"kubernetes.io/projected/788c3bf1-3291-4324-8a4c-dc06f1027f5b-kube-api-access-hcb4v\") pod \"openstackclient\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.323270 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.534877 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.541073 4946 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" podUID="788c3bf1-3291-4324-8a4c-dc06f1027f5b" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.552247 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.592932 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.656786 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config-secret\") pod \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.656873 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config\") pod \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.656905 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmhdc\" (UniqueName: \"kubernetes.io/projected/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-kube-api-access-xmhdc\") pod \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.657017 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-combined-ca-bundle\") pod \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\" (UID: \"73f2c6a5-bed7-4e90-a70d-9fea1581bcfb\") " Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.658991 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" (UID: "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.662830 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-kube-api-access-xmhdc" (OuterVolumeSpecName: "kube-api-access-xmhdc") pod "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" (UID: "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb"). InnerVolumeSpecName "kube-api-access-xmhdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.663440 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" (UID: "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.665840 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" (UID: "73f2c6a5-bed7-4e90-a70d-9fea1581bcfb"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.758959 4946 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.759272 4946 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.759286 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmhdc\" (UniqueName: \"kubernetes.io/projected/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-kube-api-access-xmhdc\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:10 crc kubenswrapper[4946]: I1203 07:13:10.759299 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:11 crc kubenswrapper[4946]: I1203 07:13:11.547496 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"788c3bf1-3291-4324-8a4c-dc06f1027f5b","Type":"ContainerStarted","Data":"c177e15f7c4d222a01139eb4976c53f0215cc04aae49a3fba5b140b7aa8123c1"} Dec 03 07:13:11 crc kubenswrapper[4946]: I1203 07:13:11.547511 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:13:11 crc kubenswrapper[4946]: I1203 07:13:11.557800 4946 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" podUID="788c3bf1-3291-4324-8a4c-dc06f1027f5b" Dec 03 07:13:11 crc kubenswrapper[4946]: I1203 07:13:11.605167 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f2c6a5-bed7-4e90-a70d-9fea1581bcfb" path="/var/lib/kubelet/pods/73f2c6a5-bed7-4e90-a70d-9fea1581bcfb/volumes" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.578073 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-8587b466c5-9tcks"] Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.579670 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.581949 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.581961 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.582657 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.609161 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-8587b466c5-9tcks"] Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.692033 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-config-data\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.692142 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-etc-swift\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.692382 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-internal-tls-certs\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.692434 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-run-httpd\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.692552 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-public-tls-certs\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.692817 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgc6c\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-kube-api-access-lgc6c\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.692950 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-log-httpd\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.692973 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-combined-ca-bundle\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.794812 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-internal-tls-certs\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.794871 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-run-httpd\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.794902 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-public-tls-certs\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.794955 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgc6c\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-kube-api-access-lgc6c\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.794993 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-log-httpd\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.795008 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-combined-ca-bundle\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.795031 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-config-data\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.795057 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-etc-swift\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.796086 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-run-httpd\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.797355 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-log-httpd\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.800700 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-internal-tls-certs\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.801087 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-public-tls-certs\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.804348 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-combined-ca-bundle\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.807168 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-config-data\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.810005 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-etc-swift\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.822198 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgc6c\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-kube-api-access-lgc6c\") pod \"swift-proxy-8587b466c5-9tcks\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:12 crc kubenswrapper[4946]: I1203 07:13:12.895473 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:13 crc kubenswrapper[4946]: I1203 07:13:13.428493 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-8587b466c5-9tcks"] Dec 03 07:13:13 crc kubenswrapper[4946]: I1203 07:13:13.571149 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8587b466c5-9tcks" event={"ID":"1e19d50d-c0ed-47b7-af00-6e8d2502f599","Type":"ContainerStarted","Data":"a4d32343ee4a198199b4fca6f200c4f71160788f8f08f17c4419345d909c1be7"} Dec 03 07:13:13 crc kubenswrapper[4946]: I1203 07:13:13.914435 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:13 crc kubenswrapper[4946]: I1203 07:13:13.915978 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="ceilometer-central-agent" containerID="cri-o://2ad123160118303cd4963f635c57fdf7b3362479e8906a77b0f999c31cdb79aa" gracePeriod=30 Dec 03 07:13:13 crc kubenswrapper[4946]: I1203 07:13:13.916058 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="sg-core" containerID="cri-o://20c5e7246f81600dfe918cfe6ae5bd0024fc7c85452e2cf955d1de4cce76c3d8" gracePeriod=30 Dec 03 07:13:13 crc kubenswrapper[4946]: I1203 07:13:13.916090 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="proxy-httpd" containerID="cri-o://b8a4e58ce6d08454436a5eca52e0c2ee12e584df13a439c59b479ea93d5a27f4" gracePeriod=30 Dec 03 07:13:13 crc kubenswrapper[4946]: I1203 07:13:13.916126 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="ceilometer-notification-agent" containerID="cri-o://9f9e599fc260887c2f3cdf78174f2cfd8105c2b2acb754d75ea473bfa9769f75" gracePeriod=30 Dec 03 07:13:13 crc kubenswrapper[4946]: I1203 07:13:13.930088 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.160:3000/\": EOF" Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.493648 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.583016 4946 generic.go:334] "Generic (PLEG): container finished" podID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerID="b8a4e58ce6d08454436a5eca52e0c2ee12e584df13a439c59b479ea93d5a27f4" exitCode=0 Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.583045 4946 generic.go:334] "Generic (PLEG): container finished" podID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerID="20c5e7246f81600dfe918cfe6ae5bd0024fc7c85452e2cf955d1de4cce76c3d8" exitCode=2 Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.583053 4946 generic.go:334] "Generic (PLEG): container finished" podID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerID="2ad123160118303cd4963f635c57fdf7b3362479e8906a77b0f999c31cdb79aa" exitCode=0 Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.583090 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerDied","Data":"b8a4e58ce6d08454436a5eca52e0c2ee12e584df13a439c59b479ea93d5a27f4"} Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.583133 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerDied","Data":"20c5e7246f81600dfe918cfe6ae5bd0024fc7c85452e2cf955d1de4cce76c3d8"} Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.583144 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerDied","Data":"2ad123160118303cd4963f635c57fdf7b3362479e8906a77b0f999c31cdb79aa"} Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.586650 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8587b466c5-9tcks" event={"ID":"1e19d50d-c0ed-47b7-af00-6e8d2502f599","Type":"ContainerStarted","Data":"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8"} Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.586688 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8587b466c5-9tcks" event={"ID":"1e19d50d-c0ed-47b7-af00-6e8d2502f599","Type":"ContainerStarted","Data":"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1"} Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.586817 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:14 crc kubenswrapper[4946]: I1203 07:13:14.586836 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:15 crc kubenswrapper[4946]: I1203 07:13:15.598449 4946 generic.go:334] "Generic (PLEG): container finished" podID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerID="9f9e599fc260887c2f3cdf78174f2cfd8105c2b2acb754d75ea473bfa9769f75" exitCode=0 Dec 03 07:13:15 crc kubenswrapper[4946]: I1203 07:13:15.605942 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerDied","Data":"9f9e599fc260887c2f3cdf78174f2cfd8105c2b2acb754d75ea473bfa9769f75"} Dec 03 07:13:17 crc kubenswrapper[4946]: I1203 07:13:17.619872 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-8587b466c5-9tcks" podStartSLOduration=5.619854955 podStartE2EDuration="5.619854955s" podCreationTimestamp="2025-12-03 07:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:13:14.62251816 +0000 UTC m=+1387.419208269" watchObservedRunningTime="2025-12-03 07:13:17.619854955 +0000 UTC m=+1390.416545064" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.124354 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.268396 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xsmv\" (UniqueName: \"kubernetes.io/projected/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-kube-api-access-7xsmv\") pod \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.268474 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-scripts\") pod \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.268533 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-config-data\") pod \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.268574 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-run-httpd\") pod \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.268612 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-combined-ca-bundle\") pod \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.268672 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-sg-core-conf-yaml\") pod \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.268772 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-log-httpd\") pod \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\" (UID: \"ccdaf2ca-24de-45e1-b757-2bee4b6572b8\") " Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.269543 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ccdaf2ca-24de-45e1-b757-2bee4b6572b8" (UID: "ccdaf2ca-24de-45e1-b757-2bee4b6572b8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.269644 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ccdaf2ca-24de-45e1-b757-2bee4b6572b8" (UID: "ccdaf2ca-24de-45e1-b757-2bee4b6572b8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.272866 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-kube-api-access-7xsmv" (OuterVolumeSpecName: "kube-api-access-7xsmv") pod "ccdaf2ca-24de-45e1-b757-2bee4b6572b8" (UID: "ccdaf2ca-24de-45e1-b757-2bee4b6572b8"). InnerVolumeSpecName "kube-api-access-7xsmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.273879 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-scripts" (OuterVolumeSpecName: "scripts") pod "ccdaf2ca-24de-45e1-b757-2bee4b6572b8" (UID: "ccdaf2ca-24de-45e1-b757-2bee4b6572b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.292414 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ccdaf2ca-24de-45e1-b757-2bee4b6572b8" (UID: "ccdaf2ca-24de-45e1-b757-2bee4b6572b8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.347921 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccdaf2ca-24de-45e1-b757-2bee4b6572b8" (UID: "ccdaf2ca-24de-45e1-b757-2bee4b6572b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.371126 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xsmv\" (UniqueName: \"kubernetes.io/projected/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-kube-api-access-7xsmv\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.371155 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.371163 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.371172 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.371180 4946 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.371188 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.384210 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-config-data" (OuterVolumeSpecName: "config-data") pod "ccdaf2ca-24de-45e1-b757-2bee4b6572b8" (UID: "ccdaf2ca-24de-45e1-b757-2bee4b6572b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.472456 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccdaf2ca-24de-45e1-b757-2bee4b6572b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.662083 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"788c3bf1-3291-4324-8a4c-dc06f1027f5b","Type":"ContainerStarted","Data":"4aaa4e05b46fa6c68df1bd9487eadf3885468fbce3baff034433d3bc2c281e0b"} Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.674768 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccdaf2ca-24de-45e1-b757-2bee4b6572b8","Type":"ContainerDied","Data":"19c0ce2d9462a3e781cd817883f21446ee19936f252c3f4d80f3e1df292633c3"} Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.674851 4946 scope.go:117] "RemoveContainer" containerID="b8a4e58ce6d08454436a5eca52e0c2ee12e584df13a439c59b479ea93d5a27f4" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.675056 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.694463 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.529847528 podStartE2EDuration="12.694443356s" podCreationTimestamp="2025-12-03 07:13:09 +0000 UTC" firstStartedPulling="2025-12-03 07:13:10.598969563 +0000 UTC m=+1383.395659682" lastFinishedPulling="2025-12-03 07:13:20.763565401 +0000 UTC m=+1393.560255510" observedRunningTime="2025-12-03 07:13:21.690812999 +0000 UTC m=+1394.487503118" watchObservedRunningTime="2025-12-03 07:13:21.694443356 +0000 UTC m=+1394.491133485" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.780396 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.783491 4946 scope.go:117] "RemoveContainer" containerID="20c5e7246f81600dfe918cfe6ae5bd0024fc7c85452e2cf955d1de4cce76c3d8" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.828995 4946 scope.go:117] "RemoveContainer" containerID="9f9e599fc260887c2f3cdf78174f2cfd8105c2b2acb754d75ea473bfa9769f75" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.829144 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.848729 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:21 crc kubenswrapper[4946]: E1203 07:13:21.849238 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="proxy-httpd" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.849260 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="proxy-httpd" Dec 03 07:13:21 crc kubenswrapper[4946]: E1203 07:13:21.850684 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="ceilometer-central-agent" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.850704 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="ceilometer-central-agent" Dec 03 07:13:21 crc kubenswrapper[4946]: E1203 07:13:21.850723 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="sg-core" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.850730 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="sg-core" Dec 03 07:13:21 crc kubenswrapper[4946]: E1203 07:13:21.850762 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="ceilometer-notification-agent" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.850769 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="ceilometer-notification-agent" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.850945 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="sg-core" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.850964 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="ceilometer-central-agent" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.850979 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="ceilometer-notification-agent" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.850989 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" containerName="proxy-httpd" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.852589 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.857089 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.860981 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.862826 4946 scope.go:117] "RemoveContainer" containerID="2ad123160118303cd4963f635c57fdf7b3362479e8906a77b0f999c31cdb79aa" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.863410 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.982682 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccz6m\" (UniqueName: \"kubernetes.io/projected/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-kube-api-access-ccz6m\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.982729 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.982777 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-config-data\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.983042 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.983290 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-log-httpd\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.983352 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-run-httpd\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:21 crc kubenswrapper[4946]: I1203 07:13:21.983388 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-scripts\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.085293 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.085363 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-config-data\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.085429 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.085482 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-log-httpd\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.085501 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-run-httpd\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.085516 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-scripts\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.085564 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccz6m\" (UniqueName: \"kubernetes.io/projected/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-kube-api-access-ccz6m\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.086789 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-log-httpd\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.088107 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-run-httpd\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.091523 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-config-data\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.091911 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.096365 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-scripts\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.107855 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.107883 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccz6m\" (UniqueName: \"kubernetes.io/projected/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-kube-api-access-ccz6m\") pod \"ceilometer-0\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.182122 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.641631 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:22 crc kubenswrapper[4946]: W1203 07:13:22.644854 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65d06492_5d60_4f6a_bb4d_a8f5dc3590ef.slice/crio-e3507f3d426c0dec78209fa88345914da67a90b5f4e20243772fdceb8b333d98 WatchSource:0}: Error finding container e3507f3d426c0dec78209fa88345914da67a90b5f4e20243772fdceb8b333d98: Status 404 returned error can't find the container with id e3507f3d426c0dec78209fa88345914da67a90b5f4e20243772fdceb8b333d98 Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.689314 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerStarted","Data":"e3507f3d426c0dec78209fa88345914da67a90b5f4e20243772fdceb8b333d98"} Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.902459 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:22 crc kubenswrapper[4946]: I1203 07:13:22.904134 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:13:23 crc kubenswrapper[4946]: I1203 07:13:23.613518 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccdaf2ca-24de-45e1-b757-2bee4b6572b8" path="/var/lib/kubelet/pods/ccdaf2ca-24de-45e1-b757-2bee4b6572b8/volumes" Dec 03 07:13:23 crc kubenswrapper[4946]: I1203 07:13:23.703677 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerStarted","Data":"ef88a8418d655973e7e657ac2fe4aacd3b23723ad93b6895d74370fd4929e119"} Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.354647 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mvhwm"] Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.357248 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.365721 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mvhwm"] Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.428714 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-utilities\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.428784 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw9w8\" (UniqueName: \"kubernetes.io/projected/a5a84cda-a324-47bb-8166-6508e691b066-kube-api-access-pw9w8\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.428820 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-catalog-content\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.518958 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.530793 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-utilities\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.531294 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-utilities\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.531432 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw9w8\" (UniqueName: \"kubernetes.io/projected/a5a84cda-a324-47bb-8166-6508e691b066-kube-api-access-pw9w8\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.531542 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-catalog-content\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.531851 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-catalog-content\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.549484 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw9w8\" (UniqueName: \"kubernetes.io/projected/a5a84cda-a324-47bb-8166-6508e691b066-kube-api-access-pw9w8\") pod \"redhat-marketplace-mvhwm\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.684262 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.713753 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerStarted","Data":"cddaba68046c69a0a9cfb7617caf0971292131d22d0bc0266b420b60c3a1284e"} Dec 03 07:13:24 crc kubenswrapper[4946]: I1203 07:13:24.713801 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerStarted","Data":"a12baf5c293b047928f7acb47d1e1a31a8e62b0f106ad7ab56fa31a96eb98785"} Dec 03 07:13:25 crc kubenswrapper[4946]: I1203 07:13:25.212073 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mvhwm"] Dec 03 07:13:25 crc kubenswrapper[4946]: I1203 07:13:25.723907 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mvhwm" event={"ID":"a5a84cda-a324-47bb-8166-6508e691b066","Type":"ContainerStarted","Data":"853f3507e7cea581091b5cd73292addb7bfdaad277368d6191cdb0d9d2ce05d2"} Dec 03 07:13:26 crc kubenswrapper[4946]: I1203 07:13:26.735532 4946 generic.go:334] "Generic (PLEG): container finished" podID="a5a84cda-a324-47bb-8166-6508e691b066" containerID="8bbf7ed4bfd39c8452df4287d669a6ba62b2a15a4a2fbbe3cb29b95583fbc7ff" exitCode=0 Dec 03 07:13:26 crc kubenswrapper[4946]: I1203 07:13:26.735577 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mvhwm" event={"ID":"a5a84cda-a324-47bb-8166-6508e691b066","Type":"ContainerDied","Data":"8bbf7ed4bfd39c8452df4287d669a6ba62b2a15a4a2fbbe3cb29b95583fbc7ff"} Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.747010 4946 generic.go:334] "Generic (PLEG): container finished" podID="a5a84cda-a324-47bb-8166-6508e691b066" containerID="16f4c8fc003db664d5e0f306fcdc880d7d0d495455f0d0af1082b3a8f6861d8c" exitCode=0 Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.747157 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mvhwm" event={"ID":"a5a84cda-a324-47bb-8166-6508e691b066","Type":"ContainerDied","Data":"16f4c8fc003db664d5e0f306fcdc880d7d0d495455f0d0af1082b3a8f6861d8c"} Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.751541 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerStarted","Data":"cb2677ef68dcf66a2fee3d9381eb7d9f9f72b50f7cd090c99beeb8ef6cdca014"} Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.751652 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="ceilometer-central-agent" containerID="cri-o://ef88a8418d655973e7e657ac2fe4aacd3b23723ad93b6895d74370fd4929e119" gracePeriod=30 Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.751717 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.751769 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="proxy-httpd" containerID="cri-o://cb2677ef68dcf66a2fee3d9381eb7d9f9f72b50f7cd090c99beeb8ef6cdca014" gracePeriod=30 Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.751804 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="sg-core" containerID="cri-o://cddaba68046c69a0a9cfb7617caf0971292131d22d0bc0266b420b60c3a1284e" gracePeriod=30 Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.751833 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="ceilometer-notification-agent" containerID="cri-o://a12baf5c293b047928f7acb47d1e1a31a8e62b0f106ad7ab56fa31a96eb98785" gracePeriod=30 Dec 03 07:13:27 crc kubenswrapper[4946]: I1203 07:13:27.794057 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.791226806 podStartE2EDuration="6.794041689s" podCreationTimestamp="2025-12-03 07:13:21 +0000 UTC" firstStartedPulling="2025-12-03 07:13:22.647260568 +0000 UTC m=+1395.443950677" lastFinishedPulling="2025-12-03 07:13:26.650075451 +0000 UTC m=+1399.446765560" observedRunningTime="2025-12-03 07:13:27.789768895 +0000 UTC m=+1400.586459004" watchObservedRunningTime="2025-12-03 07:13:27.794041689 +0000 UTC m=+1400.590731798" Dec 03 07:13:28 crc kubenswrapper[4946]: I1203 07:13:28.767192 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mvhwm" event={"ID":"a5a84cda-a324-47bb-8166-6508e691b066","Type":"ContainerStarted","Data":"02e3cd254b809591216a604bc32c72b24c11ab31b407688852a7678cf83aaefd"} Dec 03 07:13:28 crc kubenswrapper[4946]: I1203 07:13:28.777305 4946 generic.go:334] "Generic (PLEG): container finished" podID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerID="cb2677ef68dcf66a2fee3d9381eb7d9f9f72b50f7cd090c99beeb8ef6cdca014" exitCode=0 Dec 03 07:13:28 crc kubenswrapper[4946]: I1203 07:13:28.777330 4946 generic.go:334] "Generic (PLEG): container finished" podID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerID="cddaba68046c69a0a9cfb7617caf0971292131d22d0bc0266b420b60c3a1284e" exitCode=2 Dec 03 07:13:28 crc kubenswrapper[4946]: I1203 07:13:28.777339 4946 generic.go:334] "Generic (PLEG): container finished" podID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerID="a12baf5c293b047928f7acb47d1e1a31a8e62b0f106ad7ab56fa31a96eb98785" exitCode=0 Dec 03 07:13:28 crc kubenswrapper[4946]: I1203 07:13:28.777354 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerDied","Data":"cb2677ef68dcf66a2fee3d9381eb7d9f9f72b50f7cd090c99beeb8ef6cdca014"} Dec 03 07:13:28 crc kubenswrapper[4946]: I1203 07:13:28.777370 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerDied","Data":"cddaba68046c69a0a9cfb7617caf0971292131d22d0bc0266b420b60c3a1284e"} Dec 03 07:13:28 crc kubenswrapper[4946]: I1203 07:13:28.777380 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerDied","Data":"a12baf5c293b047928f7acb47d1e1a31a8e62b0f106ad7ab56fa31a96eb98785"} Dec 03 07:13:31 crc kubenswrapper[4946]: I1203 07:13:31.821690 4946 generic.go:334] "Generic (PLEG): container finished" podID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerID="ef88a8418d655973e7e657ac2fe4aacd3b23723ad93b6895d74370fd4929e119" exitCode=0 Dec 03 07:13:31 crc kubenswrapper[4946]: I1203 07:13:31.821783 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerDied","Data":"ef88a8418d655973e7e657ac2fe4aacd3b23723ad93b6895d74370fd4929e119"} Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.123809 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.160854 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mvhwm" podStartSLOduration=6.773148437 podStartE2EDuration="8.160833605s" podCreationTimestamp="2025-12-03 07:13:24 +0000 UTC" firstStartedPulling="2025-12-03 07:13:26.738858503 +0000 UTC m=+1399.535548622" lastFinishedPulling="2025-12-03 07:13:28.126543671 +0000 UTC m=+1400.923233790" observedRunningTime="2025-12-03 07:13:28.784432234 +0000 UTC m=+1401.581122363" watchObservedRunningTime="2025-12-03 07:13:32.160833605 +0000 UTC m=+1404.957523714" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.180917 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-sg-core-conf-yaml\") pod \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.181081 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-log-httpd\") pod \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.181149 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-config-data\") pod \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.181199 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-scripts\") pod \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.181243 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-run-httpd\") pod \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.181343 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccz6m\" (UniqueName: \"kubernetes.io/projected/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-kube-api-access-ccz6m\") pod \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.181408 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-combined-ca-bundle\") pod \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\" (UID: \"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef\") " Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.181787 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" (UID: "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.182026 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" (UID: "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.182525 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.182553 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.187237 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-kube-api-access-ccz6m" (OuterVolumeSpecName: "kube-api-access-ccz6m") pod "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" (UID: "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef"). InnerVolumeSpecName "kube-api-access-ccz6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.188645 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-scripts" (OuterVolumeSpecName: "scripts") pod "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" (UID: "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.214818 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" (UID: "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.271967 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" (UID: "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.284306 4946 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.284343 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.284358 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccz6m\" (UniqueName: \"kubernetes.io/projected/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-kube-api-access-ccz6m\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.284372 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.319306 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-config-data" (OuterVolumeSpecName: "config-data") pod "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" (UID: "65d06492-5d60-4f6a-bb4d-a8f5dc3590ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.386474 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.836262 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65d06492-5d60-4f6a-bb4d-a8f5dc3590ef","Type":"ContainerDied","Data":"e3507f3d426c0dec78209fa88345914da67a90b5f4e20243772fdceb8b333d98"} Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.836337 4946 scope.go:117] "RemoveContainer" containerID="cb2677ef68dcf66a2fee3d9381eb7d9f9f72b50f7cd090c99beeb8ef6cdca014" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.836357 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.884601 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.897376 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.900389 4946 scope.go:117] "RemoveContainer" containerID="cddaba68046c69a0a9cfb7617caf0971292131d22d0bc0266b420b60c3a1284e" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.906423 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:32 crc kubenswrapper[4946]: E1203 07:13:32.906770 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="sg-core" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.906785 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="sg-core" Dec 03 07:13:32 crc kubenswrapper[4946]: E1203 07:13:32.906801 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="ceilometer-notification-agent" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.906809 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="ceilometer-notification-agent" Dec 03 07:13:32 crc kubenswrapper[4946]: E1203 07:13:32.906834 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="ceilometer-central-agent" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.906842 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="ceilometer-central-agent" Dec 03 07:13:32 crc kubenswrapper[4946]: E1203 07:13:32.906854 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="proxy-httpd" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.906860 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="proxy-httpd" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.907005 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="proxy-httpd" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.907017 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="ceilometer-central-agent" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.907028 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="ceilometer-notification-agent" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.907037 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" containerName="sg-core" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.923790 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.934424 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.946487 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.956500 4946 scope.go:117] "RemoveContainer" containerID="a12baf5c293b047928f7acb47d1e1a31a8e62b0f106ad7ab56fa31a96eb98785" Dec 03 07:13:32 crc kubenswrapper[4946]: I1203 07:13:32.959340 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.000232 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.000560 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-config-data\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.000648 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.000756 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-scripts\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.000845 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-run-httpd\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.000932 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-log-httpd\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.001016 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n95vn\" (UniqueName: \"kubernetes.io/projected/2e94af06-70da-458e-8525-550541af31f3-kube-api-access-n95vn\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.103025 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-run-httpd\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.103323 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-log-httpd\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.103433 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n95vn\" (UniqueName: \"kubernetes.io/projected/2e94af06-70da-458e-8525-550541af31f3-kube-api-access-n95vn\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.103561 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.103683 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-config-data\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.103786 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.103870 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-scripts\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.105481 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-log-httpd\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.105507 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-run-httpd\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.108392 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-scripts\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.109580 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-config-data\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.121835 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.125007 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.125949 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n95vn\" (UniqueName: \"kubernetes.io/projected/2e94af06-70da-458e-8525-550541af31f3-kube-api-access-n95vn\") pod \"ceilometer-0\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.277035 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.420564 4946 scope.go:117] "RemoveContainer" containerID="ef88a8418d655973e7e657ac2fe4aacd3b23723ad93b6895d74370fd4929e119" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.606366 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d06492-5d60-4f6a-bb4d-a8f5dc3590ef" path="/var/lib/kubelet/pods/65d06492-5d60-4f6a-bb4d-a8f5dc3590ef/volumes" Dec 03 07:13:33 crc kubenswrapper[4946]: I1203 07:13:33.928442 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:34 crc kubenswrapper[4946]: I1203 07:13:34.684800 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:34 crc kubenswrapper[4946]: I1203 07:13:34.685307 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:34 crc kubenswrapper[4946]: I1203 07:13:34.740621 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:34 crc kubenswrapper[4946]: I1203 07:13:34.856896 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerStarted","Data":"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c"} Dec 03 07:13:34 crc kubenswrapper[4946]: I1203 07:13:34.856945 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerStarted","Data":"a96df235a3033f8a4b9cf258b0e7498c4feb7958e02aedfaafb7323640d90e76"} Dec 03 07:13:34 crc kubenswrapper[4946]: I1203 07:13:34.902143 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:34 crc kubenswrapper[4946]: I1203 07:13:34.983643 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mvhwm"] Dec 03 07:13:35 crc kubenswrapper[4946]: I1203 07:13:35.868590 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerStarted","Data":"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d"} Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.029570 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hn6lk"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.030758 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.038545 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hn6lk"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.129843 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-cg9b9"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.131019 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.136750 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cg9b9"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.176039 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmws4\" (UniqueName: \"kubernetes.io/projected/8489b683-060b-438e-a8eb-0615a536c35d-kube-api-access-nmws4\") pod \"nova-api-db-create-hn6lk\" (UID: \"8489b683-060b-438e-a8eb-0615a536c35d\") " pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.176170 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8489b683-060b-438e-a8eb-0615a536c35d-operator-scripts\") pod \"nova-api-db-create-hn6lk\" (UID: \"8489b683-060b-438e-a8eb-0615a536c35d\") " pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.233256 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-2e82-account-create-update-q2kxl"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.234664 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.238793 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.244436 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2e82-account-create-update-q2kxl"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.281257 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmws4\" (UniqueName: \"kubernetes.io/projected/8489b683-060b-438e-a8eb-0615a536c35d-kube-api-access-nmws4\") pod \"nova-api-db-create-hn6lk\" (UID: \"8489b683-060b-438e-a8eb-0615a536c35d\") " pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.281395 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e289880-83eb-42c7-adbe-319368792548-operator-scripts\") pod \"nova-cell0-db-create-cg9b9\" (UID: \"7e289880-83eb-42c7-adbe-319368792548\") " pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.281431 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8489b683-060b-438e-a8eb-0615a536c35d-operator-scripts\") pod \"nova-api-db-create-hn6lk\" (UID: \"8489b683-060b-438e-a8eb-0615a536c35d\") " pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.281459 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzsgd\" (UniqueName: \"kubernetes.io/projected/7e289880-83eb-42c7-adbe-319368792548-kube-api-access-wzsgd\") pod \"nova-cell0-db-create-cg9b9\" (UID: \"7e289880-83eb-42c7-adbe-319368792548\") " pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.282666 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8489b683-060b-438e-a8eb-0615a536c35d-operator-scripts\") pod \"nova-api-db-create-hn6lk\" (UID: \"8489b683-060b-438e-a8eb-0615a536c35d\") " pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.299789 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmws4\" (UniqueName: \"kubernetes.io/projected/8489b683-060b-438e-a8eb-0615a536c35d-kube-api-access-nmws4\") pod \"nova-api-db-create-hn6lk\" (UID: \"8489b683-060b-438e-a8eb-0615a536c35d\") " pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.339373 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-gtzjg"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.341915 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.349112 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.350954 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gtzjg"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.383906 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e289880-83eb-42c7-adbe-319368792548-operator-scripts\") pod \"nova-cell0-db-create-cg9b9\" (UID: \"7e289880-83eb-42c7-adbe-319368792548\") " pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.384288 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzsgd\" (UniqueName: \"kubernetes.io/projected/7e289880-83eb-42c7-adbe-319368792548-kube-api-access-wzsgd\") pod \"nova-cell0-db-create-cg9b9\" (UID: \"7e289880-83eb-42c7-adbe-319368792548\") " pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.384375 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79eeefc8-7a29-424f-b13f-40b859f7ece0-operator-scripts\") pod \"nova-api-2e82-account-create-update-q2kxl\" (UID: \"79eeefc8-7a29-424f-b13f-40b859f7ece0\") " pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.384463 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk4c2\" (UniqueName: \"kubernetes.io/projected/79eeefc8-7a29-424f-b13f-40b859f7ece0-kube-api-access-gk4c2\") pod \"nova-api-2e82-account-create-update-q2kxl\" (UID: \"79eeefc8-7a29-424f-b13f-40b859f7ece0\") " pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.384801 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e289880-83eb-42c7-adbe-319368792548-operator-scripts\") pod \"nova-cell0-db-create-cg9b9\" (UID: \"7e289880-83eb-42c7-adbe-319368792548\") " pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.403431 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzsgd\" (UniqueName: \"kubernetes.io/projected/7e289880-83eb-42c7-adbe-319368792548-kube-api-access-wzsgd\") pod \"nova-cell0-db-create-cg9b9\" (UID: \"7e289880-83eb-42c7-adbe-319368792548\") " pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.451518 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a5c7-account-create-update-qksjg"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.452286 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.452633 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.458658 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.473648 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a5c7-account-create-update-qksjg"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.487369 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be8091e1-4683-4e98-87bc-ecd638b6cbfb-operator-scripts\") pod \"nova-cell1-db-create-gtzjg\" (UID: \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\") " pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.487475 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79eeefc8-7a29-424f-b13f-40b859f7ece0-operator-scripts\") pod \"nova-api-2e82-account-create-update-q2kxl\" (UID: \"79eeefc8-7a29-424f-b13f-40b859f7ece0\") " pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.487508 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzsqn\" (UniqueName: \"kubernetes.io/projected/be8091e1-4683-4e98-87bc-ecd638b6cbfb-kube-api-access-vzsqn\") pod \"nova-cell1-db-create-gtzjg\" (UID: \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\") " pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.487572 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk4c2\" (UniqueName: \"kubernetes.io/projected/79eeefc8-7a29-424f-b13f-40b859f7ece0-kube-api-access-gk4c2\") pod \"nova-api-2e82-account-create-update-q2kxl\" (UID: \"79eeefc8-7a29-424f-b13f-40b859f7ece0\") " pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.488623 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79eeefc8-7a29-424f-b13f-40b859f7ece0-operator-scripts\") pod \"nova-api-2e82-account-create-update-q2kxl\" (UID: \"79eeefc8-7a29-424f-b13f-40b859f7ece0\") " pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.523187 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk4c2\" (UniqueName: \"kubernetes.io/projected/79eeefc8-7a29-424f-b13f-40b859f7ece0-kube-api-access-gk4c2\") pod \"nova-api-2e82-account-create-update-q2kxl\" (UID: \"79eeefc8-7a29-424f-b13f-40b859f7ece0\") " pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.558875 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.588825 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be8091e1-4683-4e98-87bc-ecd638b6cbfb-operator-scripts\") pod \"nova-cell1-db-create-gtzjg\" (UID: \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\") " pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.588901 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzsqn\" (UniqueName: \"kubernetes.io/projected/be8091e1-4683-4e98-87bc-ecd638b6cbfb-kube-api-access-vzsqn\") pod \"nova-cell1-db-create-gtzjg\" (UID: \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\") " pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.588935 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e323f7f-74ef-43bc-a54a-a061675e3dbf-operator-scripts\") pod \"nova-cell0-a5c7-account-create-update-qksjg\" (UID: \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\") " pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.588960 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llmsf\" (UniqueName: \"kubernetes.io/projected/2e323f7f-74ef-43bc-a54a-a061675e3dbf-kube-api-access-llmsf\") pod \"nova-cell0-a5c7-account-create-update-qksjg\" (UID: \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\") " pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.589638 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be8091e1-4683-4e98-87bc-ecd638b6cbfb-operator-scripts\") pod \"nova-cell1-db-create-gtzjg\" (UID: \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\") " pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.613721 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzsqn\" (UniqueName: \"kubernetes.io/projected/be8091e1-4683-4e98-87bc-ecd638b6cbfb-kube-api-access-vzsqn\") pod \"nova-cell1-db-create-gtzjg\" (UID: \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\") " pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.648761 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f06f-account-create-update-htsj6"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.649916 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.653164 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.671129 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f06f-account-create-update-htsj6"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.693596 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e323f7f-74ef-43bc-a54a-a061675e3dbf-operator-scripts\") pod \"nova-cell0-a5c7-account-create-update-qksjg\" (UID: \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\") " pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.696901 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e323f7f-74ef-43bc-a54a-a061675e3dbf-operator-scripts\") pod \"nova-cell0-a5c7-account-create-update-qksjg\" (UID: \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\") " pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.697002 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llmsf\" (UniqueName: \"kubernetes.io/projected/2e323f7f-74ef-43bc-a54a-a061675e3dbf-kube-api-access-llmsf\") pod \"nova-cell0-a5c7-account-create-update-qksjg\" (UID: \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\") " pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.723170 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llmsf\" (UniqueName: \"kubernetes.io/projected/2e323f7f-74ef-43bc-a54a-a061675e3dbf-kube-api-access-llmsf\") pod \"nova-cell0-a5c7-account-create-update-qksjg\" (UID: \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\") " pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.765504 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.775020 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.799198 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rs7k\" (UniqueName: \"kubernetes.io/projected/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-kube-api-access-9rs7k\") pod \"nova-cell1-f06f-account-create-update-htsj6\" (UID: \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\") " pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.799348 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-operator-scripts\") pod \"nova-cell1-f06f-account-create-update-htsj6\" (UID: \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\") " pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.808098 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hn6lk"] Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.882363 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hn6lk" event={"ID":"8489b683-060b-438e-a8eb-0615a536c35d","Type":"ContainerStarted","Data":"628a6ca84a8990c42de69ef65cbfc7186f0c0f7a1b0ef65c37c20b5d8fadc9f2"} Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.887794 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerStarted","Data":"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35"} Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.888043 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mvhwm" podUID="a5a84cda-a324-47bb-8166-6508e691b066" containerName="registry-server" containerID="cri-o://02e3cd254b809591216a604bc32c72b24c11ab31b407688852a7678cf83aaefd" gracePeriod=2 Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.900572 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-operator-scripts\") pod \"nova-cell1-f06f-account-create-update-htsj6\" (UID: \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\") " pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.900670 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rs7k\" (UniqueName: \"kubernetes.io/projected/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-kube-api-access-9rs7k\") pod \"nova-cell1-f06f-account-create-update-htsj6\" (UID: \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\") " pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.901469 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-operator-scripts\") pod \"nova-cell1-f06f-account-create-update-htsj6\" (UID: \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\") " pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.924648 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rs7k\" (UniqueName: \"kubernetes.io/projected/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-kube-api-access-9rs7k\") pod \"nova-cell1-f06f-account-create-update-htsj6\" (UID: \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\") " pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.969497 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:36 crc kubenswrapper[4946]: I1203 07:13:36.991327 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cg9b9"] Dec 03 07:13:37 crc kubenswrapper[4946]: W1203 07:13:37.023925 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e289880_83eb_42c7_adbe_319368792548.slice/crio-9f16099256aaf2abac43702fd6577b4b92ba6ee5a0b07bfb639feab4bd03af2d WatchSource:0}: Error finding container 9f16099256aaf2abac43702fd6577b4b92ba6ee5a0b07bfb639feab4bd03af2d: Status 404 returned error can't find the container with id 9f16099256aaf2abac43702fd6577b4b92ba6ee5a0b07bfb639feab4bd03af2d Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.095669 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2e82-account-create-update-q2kxl"] Dec 03 07:13:37 crc kubenswrapper[4946]: W1203 07:13:37.129071 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79eeefc8_7a29_424f_b13f_40b859f7ece0.slice/crio-7ea8cc746f02b165aa9588f80c36be83bcc382bcfefd974e00e5791ef5997748 WatchSource:0}: Error finding container 7ea8cc746f02b165aa9588f80c36be83bcc382bcfefd974e00e5791ef5997748: Status 404 returned error can't find the container with id 7ea8cc746f02b165aa9588f80c36be83bcc382bcfefd974e00e5791ef5997748 Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.280894 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a5c7-account-create-update-qksjg"] Dec 03 07:13:37 crc kubenswrapper[4946]: W1203 07:13:37.282848 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e323f7f_74ef_43bc_a54a_a061675e3dbf.slice/crio-94bbc7af9cfc448b975109f92513d6f46fd3a84c32b549f98ffc4fe39f77bd75 WatchSource:0}: Error finding container 94bbc7af9cfc448b975109f92513d6f46fd3a84c32b549f98ffc4fe39f77bd75: Status 404 returned error can't find the container with id 94bbc7af9cfc448b975109f92513d6f46fd3a84c32b549f98ffc4fe39f77bd75 Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.386261 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gtzjg"] Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.617526 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f06f-account-create-update-htsj6"] Dec 03 07:13:37 crc kubenswrapper[4946]: W1203 07:13:37.638485 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb24f9dfa_7c1a_4361_88d8_52b9a0ea559e.slice/crio-65c689c80e3622be5c15d478a802a2dda26e264228826c153770380b55729dbc WatchSource:0}: Error finding container 65c689c80e3622be5c15d478a802a2dda26e264228826c153770380b55729dbc: Status 404 returned error can't find the container with id 65c689c80e3622be5c15d478a802a2dda26e264228826c153770380b55729dbc Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.900133 4946 generic.go:334] "Generic (PLEG): container finished" podID="8489b683-060b-438e-a8eb-0615a536c35d" containerID="072f03744aa756782573999562fa06b8b69eba687793e806dfa9783e867eb97e" exitCode=0 Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.900441 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hn6lk" event={"ID":"8489b683-060b-438e-a8eb-0615a536c35d","Type":"ContainerDied","Data":"072f03744aa756782573999562fa06b8b69eba687793e806dfa9783e867eb97e"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.901815 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f06f-account-create-update-htsj6" event={"ID":"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e","Type":"ContainerStarted","Data":"65c689c80e3622be5c15d478a802a2dda26e264228826c153770380b55729dbc"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.902901 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gtzjg" event={"ID":"be8091e1-4683-4e98-87bc-ecd638b6cbfb","Type":"ContainerStarted","Data":"f4f5d9efdfe5afc4d55af0fcb0070d7f2a4009fb0be4c25f0c26ec8d6efc4110"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.904979 4946 generic.go:334] "Generic (PLEG): container finished" podID="79eeefc8-7a29-424f-b13f-40b859f7ece0" containerID="0dacc1b75220a156eb9ecf0c38e75527ad903c7b827431881e01db50c9cbe8e1" exitCode=0 Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.905067 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2e82-account-create-update-q2kxl" event={"ID":"79eeefc8-7a29-424f-b13f-40b859f7ece0","Type":"ContainerDied","Data":"0dacc1b75220a156eb9ecf0c38e75527ad903c7b827431881e01db50c9cbe8e1"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.905111 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2e82-account-create-update-q2kxl" event={"ID":"79eeefc8-7a29-424f-b13f-40b859f7ece0","Type":"ContainerStarted","Data":"7ea8cc746f02b165aa9588f80c36be83bcc382bcfefd974e00e5791ef5997748"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.908683 4946 generic.go:334] "Generic (PLEG): container finished" podID="7e289880-83eb-42c7-adbe-319368792548" containerID="fe14b127e3c75b219b08fa364540cfc92ff3fbff0bfffa175bcfa8f37f424f19" exitCode=0 Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.908728 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cg9b9" event={"ID":"7e289880-83eb-42c7-adbe-319368792548","Type":"ContainerDied","Data":"fe14b127e3c75b219b08fa364540cfc92ff3fbff0bfffa175bcfa8f37f424f19"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.908879 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cg9b9" event={"ID":"7e289880-83eb-42c7-adbe-319368792548","Type":"ContainerStarted","Data":"9f16099256aaf2abac43702fd6577b4b92ba6ee5a0b07bfb639feab4bd03af2d"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.914106 4946 generic.go:334] "Generic (PLEG): container finished" podID="a5a84cda-a324-47bb-8166-6508e691b066" containerID="02e3cd254b809591216a604bc32c72b24c11ab31b407688852a7678cf83aaefd" exitCode=0 Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.914177 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mvhwm" event={"ID":"a5a84cda-a324-47bb-8166-6508e691b066","Type":"ContainerDied","Data":"02e3cd254b809591216a604bc32c72b24c11ab31b407688852a7678cf83aaefd"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.914203 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mvhwm" event={"ID":"a5a84cda-a324-47bb-8166-6508e691b066","Type":"ContainerDied","Data":"853f3507e7cea581091b5cd73292addb7bfdaad277368d6191cdb0d9d2ce05d2"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.914219 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="853f3507e7cea581091b5cd73292addb7bfdaad277368d6191cdb0d9d2ce05d2" Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.918906 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerStarted","Data":"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.918944 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.918968 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.920315 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" event={"ID":"2e323f7f-74ef-43bc-a54a-a061675e3dbf","Type":"ContainerStarted","Data":"94bbc7af9cfc448b975109f92513d6f46fd3a84c32b549f98ffc4fe39f77bd75"} Dec 03 07:13:37 crc kubenswrapper[4946]: I1203 07:13:37.996307 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.882322651 podStartE2EDuration="5.996291692s" podCreationTimestamp="2025-12-03 07:13:32 +0000 UTC" firstStartedPulling="2025-12-03 07:13:33.942875787 +0000 UTC m=+1406.739565906" lastFinishedPulling="2025-12-03 07:13:37.056844838 +0000 UTC m=+1409.853534947" observedRunningTime="2025-12-03 07:13:37.988773401 +0000 UTC m=+1410.785463510" watchObservedRunningTime="2025-12-03 07:13:37.996291692 +0000 UTC m=+1410.792981801" Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.036338 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-utilities\") pod \"a5a84cda-a324-47bb-8166-6508e691b066\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.036393 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw9w8\" (UniqueName: \"kubernetes.io/projected/a5a84cda-a324-47bb-8166-6508e691b066-kube-api-access-pw9w8\") pod \"a5a84cda-a324-47bb-8166-6508e691b066\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.036591 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-catalog-content\") pod \"a5a84cda-a324-47bb-8166-6508e691b066\" (UID: \"a5a84cda-a324-47bb-8166-6508e691b066\") " Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.038463 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-utilities" (OuterVolumeSpecName: "utilities") pod "a5a84cda-a324-47bb-8166-6508e691b066" (UID: "a5a84cda-a324-47bb-8166-6508e691b066"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.054495 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5a84cda-a324-47bb-8166-6508e691b066" (UID: "a5a84cda-a324-47bb-8166-6508e691b066"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.139041 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.139073 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a84cda-a324-47bb-8166-6508e691b066-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.166077 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a84cda-a324-47bb-8166-6508e691b066-kube-api-access-pw9w8" (OuterVolumeSpecName: "kube-api-access-pw9w8") pod "a5a84cda-a324-47bb-8166-6508e691b066" (UID: "a5a84cda-a324-47bb-8166-6508e691b066"). InnerVolumeSpecName "kube-api-access-pw9w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.240132 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw9w8\" (UniqueName: \"kubernetes.io/projected/a5a84cda-a324-47bb-8166-6508e691b066-kube-api-access-pw9w8\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.930536 4946 generic.go:334] "Generic (PLEG): container finished" podID="b24f9dfa-7c1a-4361-88d8-52b9a0ea559e" containerID="86f69893a1328049356c59c5662729d8162a84f1a36a98ea3bea5e219a4226d6" exitCode=0 Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.930644 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f06f-account-create-update-htsj6" event={"ID":"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e","Type":"ContainerDied","Data":"86f69893a1328049356c59c5662729d8162a84f1a36a98ea3bea5e219a4226d6"} Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.932961 4946 generic.go:334] "Generic (PLEG): container finished" podID="be8091e1-4683-4e98-87bc-ecd638b6cbfb" containerID="d8b6ba36f5630c14206454eeddd3f93c889a0334d0f55d1e24da48d1af5af11e" exitCode=0 Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.933032 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gtzjg" event={"ID":"be8091e1-4683-4e98-87bc-ecd638b6cbfb","Type":"ContainerDied","Data":"d8b6ba36f5630c14206454eeddd3f93c889a0334d0f55d1e24da48d1af5af11e"} Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.935271 4946 generic.go:334] "Generic (PLEG): container finished" podID="2e323f7f-74ef-43bc-a54a-a061675e3dbf" containerID="29cdfda04c231794a246ad8f35d301624ae9c4ccf115d7201fda3eefc9f74be9" exitCode=0 Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.935311 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" event={"ID":"2e323f7f-74ef-43bc-a54a-a061675e3dbf","Type":"ContainerDied","Data":"29cdfda04c231794a246ad8f35d301624ae9c4ccf115d7201fda3eefc9f74be9"} Dec 03 07:13:38 crc kubenswrapper[4946]: I1203 07:13:38.935712 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mvhwm" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.008481 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mvhwm"] Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.021977 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mvhwm"] Dec 03 07:13:39 crc kubenswrapper[4946]: E1203 07:13:39.102806 4946 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5a84cda_a324_47bb_8166_6508e691b066.slice/crio-853f3507e7cea581091b5cd73292addb7bfdaad277368d6191cdb0d9d2ce05d2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5a84cda_a324_47bb_8166_6508e691b066.slice\": RecentStats: unable to find data in memory cache]" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.424683 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.431939 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.470275 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.569548 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk4c2\" (UniqueName: \"kubernetes.io/projected/79eeefc8-7a29-424f-b13f-40b859f7ece0-kube-api-access-gk4c2\") pod \"79eeefc8-7a29-424f-b13f-40b859f7ece0\" (UID: \"79eeefc8-7a29-424f-b13f-40b859f7ece0\") " Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.569591 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79eeefc8-7a29-424f-b13f-40b859f7ece0-operator-scripts\") pod \"79eeefc8-7a29-424f-b13f-40b859f7ece0\" (UID: \"79eeefc8-7a29-424f-b13f-40b859f7ece0\") " Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.569650 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmws4\" (UniqueName: \"kubernetes.io/projected/8489b683-060b-438e-a8eb-0615a536c35d-kube-api-access-nmws4\") pod \"8489b683-060b-438e-a8eb-0615a536c35d\" (UID: \"8489b683-060b-438e-a8eb-0615a536c35d\") " Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.569710 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8489b683-060b-438e-a8eb-0615a536c35d-operator-scripts\") pod \"8489b683-060b-438e-a8eb-0615a536c35d\" (UID: \"8489b683-060b-438e-a8eb-0615a536c35d\") " Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.569754 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e289880-83eb-42c7-adbe-319368792548-operator-scripts\") pod \"7e289880-83eb-42c7-adbe-319368792548\" (UID: \"7e289880-83eb-42c7-adbe-319368792548\") " Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.569837 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzsgd\" (UniqueName: \"kubernetes.io/projected/7e289880-83eb-42c7-adbe-319368792548-kube-api-access-wzsgd\") pod \"7e289880-83eb-42c7-adbe-319368792548\" (UID: \"7e289880-83eb-42c7-adbe-319368792548\") " Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.570346 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e289880-83eb-42c7-adbe-319368792548-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e289880-83eb-42c7-adbe-319368792548" (UID: "7e289880-83eb-42c7-adbe-319368792548"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.570345 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79eeefc8-7a29-424f-b13f-40b859f7ece0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79eeefc8-7a29-424f-b13f-40b859f7ece0" (UID: "79eeefc8-7a29-424f-b13f-40b859f7ece0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.570464 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8489b683-060b-438e-a8eb-0615a536c35d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8489b683-060b-438e-a8eb-0615a536c35d" (UID: "8489b683-060b-438e-a8eb-0615a536c35d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.577888 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8489b683-060b-438e-a8eb-0615a536c35d-kube-api-access-nmws4" (OuterVolumeSpecName: "kube-api-access-nmws4") pod "8489b683-060b-438e-a8eb-0615a536c35d" (UID: "8489b683-060b-438e-a8eb-0615a536c35d"). InnerVolumeSpecName "kube-api-access-nmws4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.590549 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e289880-83eb-42c7-adbe-319368792548-kube-api-access-wzsgd" (OuterVolumeSpecName: "kube-api-access-wzsgd") pod "7e289880-83eb-42c7-adbe-319368792548" (UID: "7e289880-83eb-42c7-adbe-319368792548"). InnerVolumeSpecName "kube-api-access-wzsgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.590622 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79eeefc8-7a29-424f-b13f-40b859f7ece0-kube-api-access-gk4c2" (OuterVolumeSpecName: "kube-api-access-gk4c2") pod "79eeefc8-7a29-424f-b13f-40b859f7ece0" (UID: "79eeefc8-7a29-424f-b13f-40b859f7ece0"). InnerVolumeSpecName "kube-api-access-gk4c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.604509 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a84cda-a324-47bb-8166-6508e691b066" path="/var/lib/kubelet/pods/a5a84cda-a324-47bb-8166-6508e691b066/volumes" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.671888 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e289880-83eb-42c7-adbe-319368792548-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.671923 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzsgd\" (UniqueName: \"kubernetes.io/projected/7e289880-83eb-42c7-adbe-319368792548-kube-api-access-wzsgd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.671933 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk4c2\" (UniqueName: \"kubernetes.io/projected/79eeefc8-7a29-424f-b13f-40b859f7ece0-kube-api-access-gk4c2\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.671942 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79eeefc8-7a29-424f-b13f-40b859f7ece0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.671951 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmws4\" (UniqueName: \"kubernetes.io/projected/8489b683-060b-438e-a8eb-0615a536c35d-kube-api-access-nmws4\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.671961 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8489b683-060b-438e-a8eb-0615a536c35d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.947998 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cg9b9" event={"ID":"7e289880-83eb-42c7-adbe-319368792548","Type":"ContainerDied","Data":"9f16099256aaf2abac43702fd6577b4b92ba6ee5a0b07bfb639feab4bd03af2d"} Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.948016 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cg9b9" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.948063 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f16099256aaf2abac43702fd6577b4b92ba6ee5a0b07bfb639feab4bd03af2d" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.949431 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hn6lk" event={"ID":"8489b683-060b-438e-a8eb-0615a536c35d","Type":"ContainerDied","Data":"628a6ca84a8990c42de69ef65cbfc7186f0c0f7a1b0ef65c37c20b5d8fadc9f2"} Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.949454 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="628a6ca84a8990c42de69ef65cbfc7186f0c0f7a1b0ef65c37c20b5d8fadc9f2" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.949467 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hn6lk" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.951941 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2e82-account-create-update-q2kxl" Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.952104 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2e82-account-create-update-q2kxl" event={"ID":"79eeefc8-7a29-424f-b13f-40b859f7ece0","Type":"ContainerDied","Data":"7ea8cc746f02b165aa9588f80c36be83bcc382bcfefd974e00e5791ef5997748"} Dec 03 07:13:39 crc kubenswrapper[4946]: I1203 07:13:39.952133 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ea8cc746f02b165aa9588f80c36be83bcc382bcfefd974e00e5791ef5997748" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.464066 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.471115 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.476811 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.593647 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-operator-scripts\") pod \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\" (UID: \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\") " Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.593845 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzsqn\" (UniqueName: \"kubernetes.io/projected/be8091e1-4683-4e98-87bc-ecd638b6cbfb-kube-api-access-vzsqn\") pod \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\" (UID: \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\") " Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.593902 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e323f7f-74ef-43bc-a54a-a061675e3dbf-operator-scripts\") pod \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\" (UID: \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\") " Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.593927 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llmsf\" (UniqueName: \"kubernetes.io/projected/2e323f7f-74ef-43bc-a54a-a061675e3dbf-kube-api-access-llmsf\") pod \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\" (UID: \"2e323f7f-74ef-43bc-a54a-a061675e3dbf\") " Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.593956 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be8091e1-4683-4e98-87bc-ecd638b6cbfb-operator-scripts\") pod \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\" (UID: \"be8091e1-4683-4e98-87bc-ecd638b6cbfb\") " Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.593992 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rs7k\" (UniqueName: \"kubernetes.io/projected/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-kube-api-access-9rs7k\") pod \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\" (UID: \"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e\") " Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.594210 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e323f7f-74ef-43bc-a54a-a061675e3dbf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e323f7f-74ef-43bc-a54a-a061675e3dbf" (UID: "2e323f7f-74ef-43bc-a54a-a061675e3dbf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.594414 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e323f7f-74ef-43bc-a54a-a061675e3dbf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.594571 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be8091e1-4683-4e98-87bc-ecd638b6cbfb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be8091e1-4683-4e98-87bc-ecd638b6cbfb" (UID: "be8091e1-4683-4e98-87bc-ecd638b6cbfb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.594687 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b24f9dfa-7c1a-4361-88d8-52b9a0ea559e" (UID: "b24f9dfa-7c1a-4361-88d8-52b9a0ea559e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.598516 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be8091e1-4683-4e98-87bc-ecd638b6cbfb-kube-api-access-vzsqn" (OuterVolumeSpecName: "kube-api-access-vzsqn") pod "be8091e1-4683-4e98-87bc-ecd638b6cbfb" (UID: "be8091e1-4683-4e98-87bc-ecd638b6cbfb"). InnerVolumeSpecName "kube-api-access-vzsqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.600361 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e323f7f-74ef-43bc-a54a-a061675e3dbf-kube-api-access-llmsf" (OuterVolumeSpecName: "kube-api-access-llmsf") pod "2e323f7f-74ef-43bc-a54a-a061675e3dbf" (UID: "2e323f7f-74ef-43bc-a54a-a061675e3dbf"). InnerVolumeSpecName "kube-api-access-llmsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.605863 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-kube-api-access-9rs7k" (OuterVolumeSpecName: "kube-api-access-9rs7k") pod "b24f9dfa-7c1a-4361-88d8-52b9a0ea559e" (UID: "b24f9dfa-7c1a-4361-88d8-52b9a0ea559e"). InnerVolumeSpecName "kube-api-access-9rs7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.696230 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.696261 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzsqn\" (UniqueName: \"kubernetes.io/projected/be8091e1-4683-4e98-87bc-ecd638b6cbfb-kube-api-access-vzsqn\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.696272 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llmsf\" (UniqueName: \"kubernetes.io/projected/2e323f7f-74ef-43bc-a54a-a061675e3dbf-kube-api-access-llmsf\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.696282 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be8091e1-4683-4e98-87bc-ecd638b6cbfb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.696290 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rs7k\" (UniqueName: \"kubernetes.io/projected/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e-kube-api-access-9rs7k\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.960312 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f06f-account-create-update-htsj6" event={"ID":"b24f9dfa-7c1a-4361-88d8-52b9a0ea559e","Type":"ContainerDied","Data":"65c689c80e3622be5c15d478a802a2dda26e264228826c153770380b55729dbc"} Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.960350 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65c689c80e3622be5c15d478a802a2dda26e264228826c153770380b55729dbc" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.960406 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f06f-account-create-update-htsj6" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.965024 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gtzjg" event={"ID":"be8091e1-4683-4e98-87bc-ecd638b6cbfb","Type":"ContainerDied","Data":"f4f5d9efdfe5afc4d55af0fcb0070d7f2a4009fb0be4c25f0c26ec8d6efc4110"} Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.965053 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4f5d9efdfe5afc4d55af0fcb0070d7f2a4009fb0be4c25f0c26ec8d6efc4110" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.965098 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gtzjg" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.967215 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" event={"ID":"2e323f7f-74ef-43bc-a54a-a061675e3dbf","Type":"ContainerDied","Data":"94bbc7af9cfc448b975109f92513d6f46fd3a84c32b549f98ffc4fe39f77bd75"} Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.967247 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94bbc7af9cfc448b975109f92513d6f46fd3a84c32b549f98ffc4fe39f77bd75" Dec 03 07:13:40 crc kubenswrapper[4946]: I1203 07:13:40.967295 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a5c7-account-create-update-qksjg" Dec 03 07:13:41 crc kubenswrapper[4946]: I1203 07:13:41.813470 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:41 crc kubenswrapper[4946]: I1203 07:13:41.814008 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="ceilometer-central-agent" containerID="cri-o://7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c" gracePeriod=30 Dec 03 07:13:41 crc kubenswrapper[4946]: I1203 07:13:41.814124 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="proxy-httpd" containerID="cri-o://b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493" gracePeriod=30 Dec 03 07:13:41 crc kubenswrapper[4946]: I1203 07:13:41.814159 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="sg-core" containerID="cri-o://d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35" gracePeriod=30 Dec 03 07:13:41 crc kubenswrapper[4946]: I1203 07:13:41.814190 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="ceilometer-notification-agent" containerID="cri-o://ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d" gracePeriod=30 Dec 03 07:13:41 crc kubenswrapper[4946]: I1203 07:13:41.978215 4946 generic.go:334] "Generic (PLEG): container finished" podID="2e94af06-70da-458e-8525-550541af31f3" containerID="d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35" exitCode=2 Dec 03 07:13:41 crc kubenswrapper[4946]: I1203 07:13:41.978255 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerDied","Data":"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35"} Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.538321 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.637507 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-scripts\") pod \"2e94af06-70da-458e-8525-550541af31f3\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.637555 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-sg-core-conf-yaml\") pod \"2e94af06-70da-458e-8525-550541af31f3\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.637586 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-combined-ca-bundle\") pod \"2e94af06-70da-458e-8525-550541af31f3\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.637653 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-config-data\") pod \"2e94af06-70da-458e-8525-550541af31f3\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.637769 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-log-httpd\") pod \"2e94af06-70da-458e-8525-550541af31f3\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.637790 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-run-httpd\") pod \"2e94af06-70da-458e-8525-550541af31f3\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.637862 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n95vn\" (UniqueName: \"kubernetes.io/projected/2e94af06-70da-458e-8525-550541af31f3-kube-api-access-n95vn\") pod \"2e94af06-70da-458e-8525-550541af31f3\" (UID: \"2e94af06-70da-458e-8525-550541af31f3\") " Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.638572 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e94af06-70da-458e-8525-550541af31f3" (UID: "2e94af06-70da-458e-8525-550541af31f3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.638720 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e94af06-70da-458e-8525-550541af31f3" (UID: "2e94af06-70da-458e-8525-550541af31f3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.645223 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-scripts" (OuterVolumeSpecName: "scripts") pod "2e94af06-70da-458e-8525-550541af31f3" (UID: "2e94af06-70da-458e-8525-550541af31f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.647864 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e94af06-70da-458e-8525-550541af31f3-kube-api-access-n95vn" (OuterVolumeSpecName: "kube-api-access-n95vn") pod "2e94af06-70da-458e-8525-550541af31f3" (UID: "2e94af06-70da-458e-8525-550541af31f3"). InnerVolumeSpecName "kube-api-access-n95vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.666811 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e94af06-70da-458e-8525-550541af31f3" (UID: "2e94af06-70da-458e-8525-550541af31f3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.714653 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e94af06-70da-458e-8525-550541af31f3" (UID: "2e94af06-70da-458e-8525-550541af31f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.739262 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.739294 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e94af06-70da-458e-8525-550541af31f3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.739304 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n95vn\" (UniqueName: \"kubernetes.io/projected/2e94af06-70da-458e-8525-550541af31f3-kube-api-access-n95vn\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.739313 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.739321 4946 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.739330 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.752947 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-config-data" (OuterVolumeSpecName: "config-data") pod "2e94af06-70da-458e-8525-550541af31f3" (UID: "2e94af06-70da-458e-8525-550541af31f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.840711 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e94af06-70da-458e-8525-550541af31f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990044 4946 generic.go:334] "Generic (PLEG): container finished" podID="2e94af06-70da-458e-8525-550541af31f3" containerID="b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493" exitCode=0 Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990294 4946 generic.go:334] "Generic (PLEG): container finished" podID="2e94af06-70da-458e-8525-550541af31f3" containerID="ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d" exitCode=0 Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990367 4946 generic.go:334] "Generic (PLEG): container finished" podID="2e94af06-70da-458e-8525-550541af31f3" containerID="7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c" exitCode=0 Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990438 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerDied","Data":"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493"} Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990515 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerDied","Data":"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d"} Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990573 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerDied","Data":"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c"} Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990636 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e94af06-70da-458e-8525-550541af31f3","Type":"ContainerDied","Data":"a96df235a3033f8a4b9cf258b0e7498c4feb7958e02aedfaafb7323640d90e76"} Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990699 4946 scope.go:117] "RemoveContainer" containerID="b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493" Dec 03 07:13:42 crc kubenswrapper[4946]: I1203 07:13:42.990900 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.023861 4946 scope.go:117] "RemoveContainer" containerID="d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.026092 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.037562 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.051636 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052051 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="ceilometer-central-agent" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052066 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="ceilometer-central-agent" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052083 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e289880-83eb-42c7-adbe-319368792548" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052088 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e289880-83eb-42c7-adbe-319368792548" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052099 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be8091e1-4683-4e98-87bc-ecd638b6cbfb" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052106 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="be8091e1-4683-4e98-87bc-ecd638b6cbfb" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052114 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79eeefc8-7a29-424f-b13f-40b859f7ece0" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052121 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="79eeefc8-7a29-424f-b13f-40b859f7ece0" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052142 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a84cda-a324-47bb-8166-6508e691b066" containerName="extract-content" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052149 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a84cda-a324-47bb-8166-6508e691b066" containerName="extract-content" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052159 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24f9dfa-7c1a-4361-88d8-52b9a0ea559e" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052164 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24f9dfa-7c1a-4361-88d8-52b9a0ea559e" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052175 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a84cda-a324-47bb-8166-6508e691b066" containerName="extract-utilities" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052182 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a84cda-a324-47bb-8166-6508e691b066" containerName="extract-utilities" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052192 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="ceilometer-notification-agent" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052198 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="ceilometer-notification-agent" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052207 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="sg-core" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052215 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="sg-core" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052224 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a84cda-a324-47bb-8166-6508e691b066" containerName="registry-server" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052230 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a84cda-a324-47bb-8166-6508e691b066" containerName="registry-server" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052239 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e323f7f-74ef-43bc-a54a-a061675e3dbf" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052246 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e323f7f-74ef-43bc-a54a-a061675e3dbf" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052256 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="proxy-httpd" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052263 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="proxy-httpd" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.052273 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8489b683-060b-438e-a8eb-0615a536c35d" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052279 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8489b683-060b-438e-a8eb-0615a536c35d" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052461 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a84cda-a324-47bb-8166-6508e691b066" containerName="registry-server" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052477 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="sg-core" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052484 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="79eeefc8-7a29-424f-b13f-40b859f7ece0" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052497 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="ceilometer-notification-agent" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052509 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8489b683-060b-438e-a8eb-0615a536c35d" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052518 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24f9dfa-7c1a-4361-88d8-52b9a0ea559e" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052528 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="be8091e1-4683-4e98-87bc-ecd638b6cbfb" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052540 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e289880-83eb-42c7-adbe-319368792548" containerName="mariadb-database-create" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052547 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="ceilometer-central-agent" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052555 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e323f7f-74ef-43bc-a54a-a061675e3dbf" containerName="mariadb-account-create-update" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.052563 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e94af06-70da-458e-8525-550541af31f3" containerName="proxy-httpd" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.054018 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.056930 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.057155 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.062855 4946 scope.go:117] "RemoveContainer" containerID="ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.092028 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.121660 4946 scope.go:117] "RemoveContainer" containerID="7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.142038 4946 scope.go:117] "RemoveContainer" containerID="b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.142388 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493\": container with ID starting with b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493 not found: ID does not exist" containerID="b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.142421 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493"} err="failed to get container status \"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493\": rpc error: code = NotFound desc = could not find container \"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493\": container with ID starting with b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493 not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.142440 4946 scope.go:117] "RemoveContainer" containerID="d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.142627 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35\": container with ID starting with d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35 not found: ID does not exist" containerID="d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.142645 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35"} err="failed to get container status \"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35\": rpc error: code = NotFound desc = could not find container \"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35\": container with ID starting with d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35 not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.142658 4946 scope.go:117] "RemoveContainer" containerID="ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.142855 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d\": container with ID starting with ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d not found: ID does not exist" containerID="ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.142872 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d"} err="failed to get container status \"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d\": rpc error: code = NotFound desc = could not find container \"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d\": container with ID starting with ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.142883 4946 scope.go:117] "RemoveContainer" containerID="7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c" Dec 03 07:13:43 crc kubenswrapper[4946]: E1203 07:13:43.143094 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c\": container with ID starting with 7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c not found: ID does not exist" containerID="7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.143112 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c"} err="failed to get container status \"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c\": rpc error: code = NotFound desc = could not find container \"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c\": container with ID starting with 7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.143123 4946 scope.go:117] "RemoveContainer" containerID="b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.143339 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493"} err="failed to get container status \"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493\": rpc error: code = NotFound desc = could not find container \"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493\": container with ID starting with b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493 not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.143356 4946 scope.go:117] "RemoveContainer" containerID="d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.148715 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35"} err="failed to get container status \"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35\": rpc error: code = NotFound desc = could not find container \"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35\": container with ID starting with d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35 not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.148799 4946 scope.go:117] "RemoveContainer" containerID="ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.149190 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d"} err="failed to get container status \"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d\": rpc error: code = NotFound desc = could not find container \"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d\": container with ID starting with ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.149224 4946 scope.go:117] "RemoveContainer" containerID="7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.149467 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c"} err="failed to get container status \"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c\": rpc error: code = NotFound desc = could not find container \"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c\": container with ID starting with 7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.149483 4946 scope.go:117] "RemoveContainer" containerID="b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.149640 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493"} err="failed to get container status \"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493\": rpc error: code = NotFound desc = could not find container \"b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493\": container with ID starting with b1bd2b78c0b8904b1a29edd057fa7c8d1adce091d738708f5c5204f36393f493 not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.149651 4946 scope.go:117] "RemoveContainer" containerID="d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.149854 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35"} err="failed to get container status \"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35\": rpc error: code = NotFound desc = could not find container \"d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35\": container with ID starting with d360c15f7029e0027e15fad5eb4ab9a50c42c8cf714bf4914c79a3370982ac35 not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.149865 4946 scope.go:117] "RemoveContainer" containerID="ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.150359 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-config-data\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.150378 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-log-httpd\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.150438 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.150514 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gznt\" (UniqueName: \"kubernetes.io/projected/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-kube-api-access-7gznt\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.150564 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-run-httpd\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.150843 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-scripts\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.150993 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.151924 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d"} err="failed to get container status \"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d\": rpc error: code = NotFound desc = could not find container \"ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d\": container with ID starting with ec0f81f907206f05423178c6676182854b8ef7ee4d7a54dd64f6c76f0812532d not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.152037 4946 scope.go:117] "RemoveContainer" containerID="7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.152528 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c"} err="failed to get container status \"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c\": rpc error: code = NotFound desc = could not find container \"7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c\": container with ID starting with 7fc0ed065bb5ffe23c40541572486f3785167d2902050f937ba4e8b86f41841c not found: ID does not exist" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.252470 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gznt\" (UniqueName: \"kubernetes.io/projected/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-kube-api-access-7gznt\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.252546 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-run-httpd\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.252635 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-scripts\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.252693 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.252719 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-config-data\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.252755 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-log-httpd\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.252786 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.253665 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-run-httpd\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.253721 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-log-httpd\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.258984 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.259229 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-scripts\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.259572 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-config-data\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.259635 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.273964 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gznt\" (UniqueName: \"kubernetes.io/projected/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-kube-api-access-7gznt\") pod \"ceilometer-0\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.298292 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.299186 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-httpd" containerID="cri-o://7664bde7ec828564ba96e4d3d5111e746146cbd7db9fad042b1af00c47996a16" gracePeriod=30 Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.299712 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-log" containerID="cri-o://6e66bc9d1b8394a530308eae9d3fb3ffc5e8f1512eef972c0d4c3b73a556409d" gracePeriod=30 Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.385898 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.603088 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e94af06-70da-458e-8525-550541af31f3" path="/var/lib/kubelet/pods/2e94af06-70da-458e-8525-550541af31f3/volumes" Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.906685 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:43 crc kubenswrapper[4946]: I1203 07:13:43.913899 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:13:44 crc kubenswrapper[4946]: I1203 07:13:44.000718 4946 generic.go:334] "Generic (PLEG): container finished" podID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerID="6e66bc9d1b8394a530308eae9d3fb3ffc5e8f1512eef972c0d4c3b73a556409d" exitCode=143 Dec 03 07:13:44 crc kubenswrapper[4946]: I1203 07:13:44.000800 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7","Type":"ContainerDied","Data":"6e66bc9d1b8394a530308eae9d3fb3ffc5e8f1512eef972c0d4c3b73a556409d"} Dec 03 07:13:44 crc kubenswrapper[4946]: I1203 07:13:44.004139 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerStarted","Data":"a15cb40f144382ebea34501d1a0b71045ae9095270fd853c8c82a619c3bb8ade"} Dec 03 07:13:44 crc kubenswrapper[4946]: I1203 07:13:44.141065 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:13:44 crc kubenswrapper[4946]: I1203 07:13:44.141340 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerName="glance-log" containerID="cri-o://3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517" gracePeriod=30 Dec 03 07:13:44 crc kubenswrapper[4946]: I1203 07:13:44.141446 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerName="glance-httpd" containerID="cri-o://3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1" gracePeriod=30 Dec 03 07:13:45 crc kubenswrapper[4946]: I1203 07:13:45.014643 4946 generic.go:334] "Generic (PLEG): container finished" podID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerID="3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517" exitCode=143 Dec 03 07:13:45 crc kubenswrapper[4946]: I1203 07:13:45.014794 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0858d6c7-5e35-4c61-b0fd-70cd101c0498","Type":"ContainerDied","Data":"3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517"} Dec 03 07:13:45 crc kubenswrapper[4946]: I1203 07:13:45.017159 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerStarted","Data":"3094708632df5b159d2bf6133a4f3ee720f6efc93a22e2b6ef2a2b1ca81637fd"} Dec 03 07:13:45 crc kubenswrapper[4946]: I1203 07:13:45.353356 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.027217 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerStarted","Data":"bdd42e0495c25db0bfd7a6bee417a5e00de9e626919b6c07b7165d2e59deff1e"} Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.027564 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerStarted","Data":"315daa8d7809bce784559041f6d32310303fe1c049ed5f9e01eed7bf24ba6320"} Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.495378 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.145:9292/healthcheck\": dial tcp 10.217.0.145:9292: connect: connection refused" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.495950 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.145:9292/healthcheck\": dial tcp 10.217.0.145:9292: connect: connection refused" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.754940 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xp9sr"] Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.764770 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.766751 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.767056 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.768149 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8j4l2" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.811446 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xp9sr"] Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.920835 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-scripts\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.920928 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.920967 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf9fm\" (UniqueName: \"kubernetes.io/projected/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-kube-api-access-lf9fm\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:46 crc kubenswrapper[4946]: I1203 07:13:46.920987 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.022861 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.022926 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf9fm\" (UniqueName: \"kubernetes.io/projected/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-kube-api-access-lf9fm\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.022953 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.023028 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-scripts\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.027310 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.027839 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.037843 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-scripts\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.043545 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf9fm\" (UniqueName: \"kubernetes.io/projected/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-kube-api-access-lf9fm\") pod \"nova-cell0-conductor-db-sync-xp9sr\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.095589 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.106156 4946 generic.go:334] "Generic (PLEG): container finished" podID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerID="7664bde7ec828564ba96e4d3d5111e746146cbd7db9fad042b1af00c47996a16" exitCode=0 Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.106202 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7","Type":"ContainerDied","Data":"7664bde7ec828564ba96e4d3d5111e746146cbd7db9fad042b1af00c47996a16"} Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.113789 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.239536 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.239893 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-logs\") pod \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.239924 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-scripts\") pod \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.240024 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-combined-ca-bundle\") pod \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.240065 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-config-data\") pod \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.240082 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7zmw\" (UniqueName: \"kubernetes.io/projected/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-kube-api-access-b7zmw\") pod \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.240159 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-httpd-run\") pod \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.240194 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-public-tls-certs\") pod \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\" (UID: \"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7\") " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.240312 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-logs" (OuterVolumeSpecName: "logs") pod "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" (UID: "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.240526 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.243456 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" (UID: "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.243651 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" (UID: "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.243677 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-kube-api-access-b7zmw" (OuterVolumeSpecName: "kube-api-access-b7zmw") pod "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" (UID: "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7"). InnerVolumeSpecName "kube-api-access-b7zmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.246912 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-scripts" (OuterVolumeSpecName: "scripts") pod "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" (UID: "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.300839 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" (UID: "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.326821 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-config-data" (OuterVolumeSpecName: "config-data") pod "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" (UID: "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.338563 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" (UID: "74ecf325-8a5b-4bd6-bdd6-d9c935afafc7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.342422 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.342447 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.342456 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7zmw\" (UniqueName: \"kubernetes.io/projected/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-kube-api-access-b7zmw\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.342467 4946 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.342475 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.342506 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.342514 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.361846 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.444089 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.629807 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xp9sr"] Dec 03 07:13:47 crc kubenswrapper[4946]: I1203 07:13:47.863323 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.053220 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hb4k\" (UniqueName: \"kubernetes.io/projected/0858d6c7-5e35-4c61-b0fd-70cd101c0498-kube-api-access-8hb4k\") pod \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.053262 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-config-data\") pod \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.053287 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.053307 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-internal-tls-certs\") pod \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.053336 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-logs\") pod \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.053386 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-combined-ca-bundle\") pod \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.053433 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-scripts\") pod \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.053452 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-httpd-run\") pod \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\" (UID: \"0858d6c7-5e35-4c61-b0fd-70cd101c0498\") " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.054146 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0858d6c7-5e35-4c61-b0fd-70cd101c0498" (UID: "0858d6c7-5e35-4c61-b0fd-70cd101c0498"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.055468 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-logs" (OuterVolumeSpecName: "logs") pod "0858d6c7-5e35-4c61-b0fd-70cd101c0498" (UID: "0858d6c7-5e35-4c61-b0fd-70cd101c0498"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.060078 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "0858d6c7-5e35-4c61-b0fd-70cd101c0498" (UID: "0858d6c7-5e35-4c61-b0fd-70cd101c0498"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.060128 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0858d6c7-5e35-4c61-b0fd-70cd101c0498-kube-api-access-8hb4k" (OuterVolumeSpecName: "kube-api-access-8hb4k") pod "0858d6c7-5e35-4c61-b0fd-70cd101c0498" (UID: "0858d6c7-5e35-4c61-b0fd-70cd101c0498"). InnerVolumeSpecName "kube-api-access-8hb4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.060458 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-scripts" (OuterVolumeSpecName: "scripts") pod "0858d6c7-5e35-4c61-b0fd-70cd101c0498" (UID: "0858d6c7-5e35-4c61-b0fd-70cd101c0498"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.100027 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0858d6c7-5e35-4c61-b0fd-70cd101c0498" (UID: "0858d6c7-5e35-4c61-b0fd-70cd101c0498"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.106930 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-config-data" (OuterVolumeSpecName: "config-data") pod "0858d6c7-5e35-4c61-b0fd-70cd101c0498" (UID: "0858d6c7-5e35-4c61-b0fd-70cd101c0498"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.117716 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" event={"ID":"c6109109-dfb1-4b8c-86c7-94d7fe00fed1","Type":"ContainerStarted","Data":"22dcb657c0ebe3eba25472b4a4409735ccf6760c22153bf82e198f2f2cb2b0a7"} Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.119696 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0858d6c7-5e35-4c61-b0fd-70cd101c0498" (UID: "0858d6c7-5e35-4c61-b0fd-70cd101c0498"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.121315 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74ecf325-8a5b-4bd6-bdd6-d9c935afafc7","Type":"ContainerDied","Data":"0cc93719d4efd02d1e0636175b9f92adb30f05187af5552a57a9578619eccea9"} Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.121359 4946 scope.go:117] "RemoveContainer" containerID="7664bde7ec828564ba96e4d3d5111e746146cbd7db9fad042b1af00c47996a16" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.121423 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.128465 4946 generic.go:334] "Generic (PLEG): container finished" podID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerID="3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1" exitCode=0 Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.128522 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0858d6c7-5e35-4c61-b0fd-70cd101c0498","Type":"ContainerDied","Data":"3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1"} Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.128544 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0858d6c7-5e35-4c61-b0fd-70cd101c0498","Type":"ContainerDied","Data":"cacc1f5df15dfb52ff77dd2b14a4908d4354bea6ea2f5f7f8119443a9dd73da2"} Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.128631 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.133545 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerStarted","Data":"a8fcde969d6587f94b75052c2082af4c747c7ae565921025cb63af54faad0ca2"} Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.134004 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.134053 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="sg-core" containerID="cri-o://bdd42e0495c25db0bfd7a6bee417a5e00de9e626919b6c07b7165d2e59deff1e" gracePeriod=30 Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.134119 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="proxy-httpd" containerID="cri-o://a8fcde969d6587f94b75052c2082af4c747c7ae565921025cb63af54faad0ca2" gracePeriod=30 Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.134160 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="ceilometer-notification-agent" containerID="cri-o://315daa8d7809bce784559041f6d32310303fe1c049ed5f9e01eed7bf24ba6320" gracePeriod=30 Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.134220 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="ceilometer-central-agent" containerID="cri-o://3094708632df5b159d2bf6133a4f3ee720f6efc93a22e2b6ef2a2b1ca81637fd" gracePeriod=30 Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.154778 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.154804 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.154816 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.154823 4946 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0858d6c7-5e35-4c61-b0fd-70cd101c0498-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.154832 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hb4k\" (UniqueName: \"kubernetes.io/projected/0858d6c7-5e35-4c61-b0fd-70cd101c0498-kube-api-access-8hb4k\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.154840 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.154862 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.154871 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0858d6c7-5e35-4c61-b0fd-70cd101c0498-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.160454 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.173225 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.189369 4946 scope.go:117] "RemoveContainer" containerID="6e66bc9d1b8394a530308eae9d3fb3ffc5e8f1512eef972c0d4c3b73a556409d" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.189636 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:13:48 crc kubenswrapper[4946]: E1203 07:13:48.190170 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerName="glance-log" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.190247 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerName="glance-log" Dec 03 07:13:48 crc kubenswrapper[4946]: E1203 07:13:48.190325 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-httpd" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.190376 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-httpd" Dec 03 07:13:48 crc kubenswrapper[4946]: E1203 07:13:48.190441 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerName="glance-httpd" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.190490 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerName="glance-httpd" Dec 03 07:13:48 crc kubenswrapper[4946]: E1203 07:13:48.190552 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-log" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.190602 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-log" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.190843 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerName="glance-log" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.190918 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-httpd" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.190981 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" containerName="glance-httpd" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.191039 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" containerName="glance-log" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.191942 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.195963 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.196135 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.196264 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.196392 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bk8hz" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.197465 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.199317 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.2016859540000002 podStartE2EDuration="5.199305106s" podCreationTimestamp="2025-12-03 07:13:43 +0000 UTC" firstStartedPulling="2025-12-03 07:13:43.913273547 +0000 UTC m=+1416.709963696" lastFinishedPulling="2025-12-03 07:13:46.910892739 +0000 UTC m=+1419.707582848" observedRunningTime="2025-12-03 07:13:48.16499796 +0000 UTC m=+1420.961688069" watchObservedRunningTime="2025-12-03 07:13:48.199305106 +0000 UTC m=+1420.995995215" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.226398 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.236789 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.245930 4946 scope.go:117] "RemoveContainer" containerID="3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.257160 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.278078 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.278900 4946 scope.go:117] "RemoveContainer" containerID="3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.287811 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.290557 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.295650 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.295914 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.323375 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.326265 4946 scope.go:117] "RemoveContainer" containerID="3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1" Dec 03 07:13:48 crc kubenswrapper[4946]: E1203 07:13:48.327508 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1\": container with ID starting with 3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1 not found: ID does not exist" containerID="3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.327792 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1"} err="failed to get container status \"3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1\": rpc error: code = NotFound desc = could not find container \"3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1\": container with ID starting with 3e665087b926a0c59ceb1632540977b8470b400282a0919a96d3b52bef0955f1 not found: ID does not exist" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.328030 4946 scope.go:117] "RemoveContainer" containerID="3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517" Dec 03 07:13:48 crc kubenswrapper[4946]: E1203 07:13:48.328439 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517\": container with ID starting with 3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517 not found: ID does not exist" containerID="3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.328478 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517"} err="failed to get container status \"3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517\": rpc error: code = NotFound desc = could not find container \"3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517\": container with ID starting with 3658e159af88ae2cc9e426d26a94e1b75e9bece6760a36e9b34e91faa38a0517 not found: ID does not exist" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.358965 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.358997 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.359052 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-logs\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.359069 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.359109 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-scripts\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.359125 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.359279 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjx9b\" (UniqueName: \"kubernetes.io/projected/bd1d9c96-a0f4-43db-a686-2b320c1948ae-kube-api-access-wjx9b\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.359492 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-config-data\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461043 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-logs\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461080 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461111 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-scripts\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461137 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461177 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461202 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjx9b\" (UniqueName: \"kubernetes.io/projected/bd1d9c96-a0f4-43db-a686-2b320c1948ae-kube-api-access-wjx9b\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461226 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461272 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-config-data\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461287 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461314 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-logs\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461337 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461355 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461383 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461402 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461429 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461446 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6mwj\" (UniqueName: \"kubernetes.io/projected/7758850c-e98d-4bb6-bd84-31584a9bcdb6-kube-api-access-t6mwj\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.461522 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-logs\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.462150 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.462519 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.467557 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-scripts\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.467589 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.467954 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-config-data\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.469252 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.479463 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjx9b\" (UniqueName: \"kubernetes.io/projected/bd1d9c96-a0f4-43db-a686-2b320c1948ae-kube-api-access-wjx9b\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.487089 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.536259 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.562953 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-logs\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.563099 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.563179 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.563295 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.563382 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6mwj\" (UniqueName: \"kubernetes.io/projected/7758850c-e98d-4bb6-bd84-31584a9bcdb6-kube-api-access-t6mwj\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.563442 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.563381 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-logs\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.563730 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.563887 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.564041 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.564241 4946 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.570712 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.570773 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.570831 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.571097 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.581341 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6mwj\" (UniqueName: \"kubernetes.io/projected/7758850c-e98d-4bb6-bd84-31584a9bcdb6-kube-api-access-t6mwj\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.590517 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " pod="openstack/glance-default-internal-api-0" Dec 03 07:13:48 crc kubenswrapper[4946]: I1203 07:13:48.622523 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.147887 4946 generic.go:334] "Generic (PLEG): container finished" podID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerID="a8fcde969d6587f94b75052c2082af4c747c7ae565921025cb63af54faad0ca2" exitCode=0 Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.148181 4946 generic.go:334] "Generic (PLEG): container finished" podID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerID="bdd42e0495c25db0bfd7a6bee417a5e00de9e626919b6c07b7165d2e59deff1e" exitCode=2 Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.148191 4946 generic.go:334] "Generic (PLEG): container finished" podID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerID="315daa8d7809bce784559041f6d32310303fe1c049ed5f9e01eed7bf24ba6320" exitCode=0 Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.147939 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerDied","Data":"a8fcde969d6587f94b75052c2082af4c747c7ae565921025cb63af54faad0ca2"} Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.148224 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerDied","Data":"bdd42e0495c25db0bfd7a6bee417a5e00de9e626919b6c07b7165d2e59deff1e"} Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.148238 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerDied","Data":"315daa8d7809bce784559041f6d32310303fe1c049ed5f9e01eed7bf24ba6320"} Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.190325 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.291881 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.606821 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0858d6c7-5e35-4c61-b0fd-70cd101c0498" path="/var/lib/kubelet/pods/0858d6c7-5e35-4c61-b0fd-70cd101c0498/volumes" Dec 03 07:13:49 crc kubenswrapper[4946]: I1203 07:13:49.607916 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74ecf325-8a5b-4bd6-bdd6-d9c935afafc7" path="/var/lib/kubelet/pods/74ecf325-8a5b-4bd6-bdd6-d9c935afafc7/volumes" Dec 03 07:13:50 crc kubenswrapper[4946]: I1203 07:13:50.167632 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7758850c-e98d-4bb6-bd84-31584a9bcdb6","Type":"ContainerStarted","Data":"33b47ed2f51fb1905563dad28b7eecdde497b32b4bd342ccb3247ecc1b377e2f"} Dec 03 07:13:50 crc kubenswrapper[4946]: I1203 07:13:50.168033 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7758850c-e98d-4bb6-bd84-31584a9bcdb6","Type":"ContainerStarted","Data":"925234f7444f86715154173321a353e7552589fab9828c5f28ecc5539d42fb86"} Dec 03 07:13:50 crc kubenswrapper[4946]: I1203 07:13:50.170896 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bd1d9c96-a0f4-43db-a686-2b320c1948ae","Type":"ContainerStarted","Data":"86788e20897ae74696026dd8776710b0bda123a3f5dc62f4eebb35926ddd9811"} Dec 03 07:13:50 crc kubenswrapper[4946]: I1203 07:13:50.170926 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bd1d9c96-a0f4-43db-a686-2b320c1948ae","Type":"ContainerStarted","Data":"3dce7a4effecb50c6bfddb3992c650827125bd582d3634d4d3e8261ace1d7b4c"} Dec 03 07:13:51 crc kubenswrapper[4946]: I1203 07:13:51.184354 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bd1d9c96-a0f4-43db-a686-2b320c1948ae","Type":"ContainerStarted","Data":"6c046b12932defb2b1b37ab4cba5b712c6fb5e6d71e6a3ef814f1586ed8b402d"} Dec 03 07:13:51 crc kubenswrapper[4946]: I1203 07:13:51.187347 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7758850c-e98d-4bb6-bd84-31584a9bcdb6","Type":"ContainerStarted","Data":"7289e3069aa5ae14b8fa7cb3805ddb5abba4f40160f9caf5ebdf388dff4bf65a"} Dec 03 07:13:51 crc kubenswrapper[4946]: I1203 07:13:51.190350 4946 generic.go:334] "Generic (PLEG): container finished" podID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerID="3094708632df5b159d2bf6133a4f3ee720f6efc93a22e2b6ef2a2b1ca81637fd" exitCode=0 Dec 03 07:13:51 crc kubenswrapper[4946]: I1203 07:13:51.190386 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerDied","Data":"3094708632df5b159d2bf6133a4f3ee720f6efc93a22e2b6ef2a2b1ca81637fd"} Dec 03 07:13:51 crc kubenswrapper[4946]: I1203 07:13:51.219495 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.219475641 podStartE2EDuration="3.219475641s" podCreationTimestamp="2025-12-03 07:13:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:13:51.209131495 +0000 UTC m=+1424.005821604" watchObservedRunningTime="2025-12-03 07:13:51.219475641 +0000 UTC m=+1424.016165760" Dec 03 07:13:51 crc kubenswrapper[4946]: I1203 07:13:51.244039 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.244012896 podStartE2EDuration="3.244012896s" podCreationTimestamp="2025-12-03 07:13:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:13:51.232461028 +0000 UTC m=+1424.029151147" watchObservedRunningTime="2025-12-03 07:13:51.244012896 +0000 UTC m=+1424.040703005" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.250282 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a27059b6-10a4-4f4d-b513-fbfb3a6664c3","Type":"ContainerDied","Data":"a15cb40f144382ebea34501d1a0b71045ae9095270fd853c8c82a619c3bb8ade"} Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.250697 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a15cb40f144382ebea34501d1a0b71045ae9095270fd853c8c82a619c3bb8ade" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.287133 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.394310 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gznt\" (UniqueName: \"kubernetes.io/projected/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-kube-api-access-7gznt\") pod \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.394412 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-run-httpd\") pod \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.394482 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-combined-ca-bundle\") pod \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.394572 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-config-data\") pod \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.394837 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a27059b6-10a4-4f4d-b513-fbfb3a6664c3" (UID: "a27059b6-10a4-4f4d-b513-fbfb3a6664c3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.394892 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-sg-core-conf-yaml\") pod \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.395035 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-scripts\") pod \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.395093 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-log-httpd\") pod \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\" (UID: \"a27059b6-10a4-4f4d-b513-fbfb3a6664c3\") " Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.395780 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a27059b6-10a4-4f4d-b513-fbfb3a6664c3" (UID: "a27059b6-10a4-4f4d-b513-fbfb3a6664c3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.396218 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.396274 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.400724 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-scripts" (OuterVolumeSpecName: "scripts") pod "a27059b6-10a4-4f4d-b513-fbfb3a6664c3" (UID: "a27059b6-10a4-4f4d-b513-fbfb3a6664c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.401020 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-kube-api-access-7gznt" (OuterVolumeSpecName: "kube-api-access-7gznt") pod "a27059b6-10a4-4f4d-b513-fbfb3a6664c3" (UID: "a27059b6-10a4-4f4d-b513-fbfb3a6664c3"). InnerVolumeSpecName "kube-api-access-7gznt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.418952 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a27059b6-10a4-4f4d-b513-fbfb3a6664c3" (UID: "a27059b6-10a4-4f4d-b513-fbfb3a6664c3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.481348 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a27059b6-10a4-4f4d-b513-fbfb3a6664c3" (UID: "a27059b6-10a4-4f4d-b513-fbfb3a6664c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.494662 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-config-data" (OuterVolumeSpecName: "config-data") pod "a27059b6-10a4-4f4d-b513-fbfb3a6664c3" (UID: "a27059b6-10a4-4f4d-b513-fbfb3a6664c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.497705 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.497758 4946 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.497770 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.497779 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gznt\" (UniqueName: \"kubernetes.io/projected/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-kube-api-access-7gznt\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:55 crc kubenswrapper[4946]: I1203 07:13:55.497787 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a27059b6-10a4-4f4d-b513-fbfb3a6664c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.264190 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.265904 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" event={"ID":"c6109109-dfb1-4b8c-86c7-94d7fe00fed1","Type":"ContainerStarted","Data":"985f669d8f4553a11987acfa7117c970727e75f62281c757c43798f37b6ae8b7"} Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.294907 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" podStartSLOduration=2.852839034 podStartE2EDuration="10.294882196s" podCreationTimestamp="2025-12-03 07:13:46 +0000 UTC" firstStartedPulling="2025-12-03 07:13:47.694322487 +0000 UTC m=+1420.491012596" lastFinishedPulling="2025-12-03 07:13:55.136365639 +0000 UTC m=+1427.933055758" observedRunningTime="2025-12-03 07:13:56.287881219 +0000 UTC m=+1429.084571398" watchObservedRunningTime="2025-12-03 07:13:56.294882196 +0000 UTC m=+1429.091572345" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.341573 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.357389 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.367810 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:56 crc kubenswrapper[4946]: E1203 07:13:56.368379 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="ceilometer-notification-agent" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.368398 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="ceilometer-notification-agent" Dec 03 07:13:56 crc kubenswrapper[4946]: E1203 07:13:56.368419 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="proxy-httpd" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.368429 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="proxy-httpd" Dec 03 07:13:56 crc kubenswrapper[4946]: E1203 07:13:56.368445 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="sg-core" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.368452 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="sg-core" Dec 03 07:13:56 crc kubenswrapper[4946]: E1203 07:13:56.368471 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="ceilometer-central-agent" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.368479 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="ceilometer-central-agent" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.368819 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="sg-core" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.368863 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="ceilometer-central-agent" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.368883 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="proxy-httpd" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.368904 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" containerName="ceilometer-notification-agent" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.371250 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.374197 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.374217 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.379287 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.432819 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-run-httpd\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.432894 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-config-data\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.432920 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-scripts\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.432963 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.432992 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-log-httpd\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.433021 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdx5v\" (UniqueName: \"kubernetes.io/projected/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-kube-api-access-mdx5v\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.433064 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.534883 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.535276 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-log-httpd\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.535402 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdx5v\" (UniqueName: \"kubernetes.io/projected/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-kube-api-access-mdx5v\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.535565 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.535693 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-run-httpd\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.535853 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-config-data\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.535968 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-scripts\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.535747 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-log-httpd\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.536185 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-run-httpd\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.540962 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-config-data\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.541148 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-scripts\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.542036 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.551255 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.558258 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdx5v\" (UniqueName: \"kubernetes.io/projected/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-kube-api-access-mdx5v\") pod \"ceilometer-0\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " pod="openstack/ceilometer-0" Dec 03 07:13:56 crc kubenswrapper[4946]: I1203 07:13:56.691224 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:13:57 crc kubenswrapper[4946]: I1203 07:13:57.154807 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:13:57 crc kubenswrapper[4946]: W1203 07:13:57.156087 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecfb79fe_51e7_491b_8e36_4aff7cbe2a8d.slice/crio-04e9927180229d82e4dfbdc8c049b094f8acfcdfb5729d166858c439350fad86 WatchSource:0}: Error finding container 04e9927180229d82e4dfbdc8c049b094f8acfcdfb5729d166858c439350fad86: Status 404 returned error can't find the container with id 04e9927180229d82e4dfbdc8c049b094f8acfcdfb5729d166858c439350fad86 Dec 03 07:13:57 crc kubenswrapper[4946]: I1203 07:13:57.277249 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerStarted","Data":"04e9927180229d82e4dfbdc8c049b094f8acfcdfb5729d166858c439350fad86"} Dec 03 07:13:57 crc kubenswrapper[4946]: I1203 07:13:57.609150 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a27059b6-10a4-4f4d-b513-fbfb3a6664c3" path="/var/lib/kubelet/pods/a27059b6-10a4-4f4d-b513-fbfb3a6664c3/volumes" Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.287758 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerStarted","Data":"e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3"} Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.538808 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.539184 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.591932 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.593836 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.630185 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.631331 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.672387 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:58 crc kubenswrapper[4946]: I1203 07:13:58.684708 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:59 crc kubenswrapper[4946]: I1203 07:13:59.297540 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerStarted","Data":"d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5"} Dec 03 07:13:59 crc kubenswrapper[4946]: I1203 07:13:59.298046 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:59 crc kubenswrapper[4946]: I1203 07:13:59.298063 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 07:13:59 crc kubenswrapper[4946]: I1203 07:13:59.298073 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 07:13:59 crc kubenswrapper[4946]: I1203 07:13:59.298082 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 07:14:01 crc kubenswrapper[4946]: I1203 07:14:01.317144 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerStarted","Data":"0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85"} Dec 03 07:14:01 crc kubenswrapper[4946]: I1203 07:14:01.555476 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 07:14:01 crc kubenswrapper[4946]: I1203 07:14:01.555597 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 07:14:01 crc kubenswrapper[4946]: I1203 07:14:01.555986 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 07:14:01 crc kubenswrapper[4946]: I1203 07:14:01.556094 4946 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 07:14:01 crc kubenswrapper[4946]: I1203 07:14:01.706539 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 07:14:01 crc kubenswrapper[4946]: I1203 07:14:01.785539 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 07:14:03 crc kubenswrapper[4946]: I1203 07:14:03.346557 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerStarted","Data":"3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c"} Dec 03 07:14:03 crc kubenswrapper[4946]: I1203 07:14:03.346948 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 07:14:03 crc kubenswrapper[4946]: I1203 07:14:03.384003 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.26949195 podStartE2EDuration="7.38398501s" podCreationTimestamp="2025-12-03 07:13:56 +0000 UTC" firstStartedPulling="2025-12-03 07:13:57.159936742 +0000 UTC m=+1429.956626851" lastFinishedPulling="2025-12-03 07:14:02.274429792 +0000 UTC m=+1435.071119911" observedRunningTime="2025-12-03 07:14:03.374706242 +0000 UTC m=+1436.171396351" watchObservedRunningTime="2025-12-03 07:14:03.38398501 +0000 UTC m=+1436.180675109" Dec 03 07:14:07 crc kubenswrapper[4946]: I1203 07:14:07.386220 4946 generic.go:334] "Generic (PLEG): container finished" podID="c6109109-dfb1-4b8c-86c7-94d7fe00fed1" containerID="985f669d8f4553a11987acfa7117c970727e75f62281c757c43798f37b6ae8b7" exitCode=0 Dec 03 07:14:07 crc kubenswrapper[4946]: I1203 07:14:07.386317 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" event={"ID":"c6109109-dfb1-4b8c-86c7-94d7fe00fed1","Type":"ContainerDied","Data":"985f669d8f4553a11987acfa7117c970727e75f62281c757c43798f37b6ae8b7"} Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.766549 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.877936 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-scripts\") pod \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.878132 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data\") pod \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.878205 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-combined-ca-bundle\") pod \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.878300 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf9fm\" (UniqueName: \"kubernetes.io/projected/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-kube-api-access-lf9fm\") pod \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.884806 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-kube-api-access-lf9fm" (OuterVolumeSpecName: "kube-api-access-lf9fm") pod "c6109109-dfb1-4b8c-86c7-94d7fe00fed1" (UID: "c6109109-dfb1-4b8c-86c7-94d7fe00fed1"). InnerVolumeSpecName "kube-api-access-lf9fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.886337 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-scripts" (OuterVolumeSpecName: "scripts") pod "c6109109-dfb1-4b8c-86c7-94d7fe00fed1" (UID: "c6109109-dfb1-4b8c-86c7-94d7fe00fed1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:08 crc kubenswrapper[4946]: E1203 07:14:08.907064 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data podName:c6109109-dfb1-4b8c-86c7-94d7fe00fed1 nodeName:}" failed. No retries permitted until 2025-12-03 07:14:09.407034322 +0000 UTC m=+1442.203724431 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data") pod "c6109109-dfb1-4b8c-86c7-94d7fe00fed1" (UID: "c6109109-dfb1-4b8c-86c7-94d7fe00fed1") : error deleting /var/lib/kubelet/pods/c6109109-dfb1-4b8c-86c7-94d7fe00fed1/volume-subpaths: remove /var/lib/kubelet/pods/c6109109-dfb1-4b8c-86c7-94d7fe00fed1/volume-subpaths: no such file or directory Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.912054 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6109109-dfb1-4b8c-86c7-94d7fe00fed1" (UID: "c6109109-dfb1-4b8c-86c7-94d7fe00fed1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.981178 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.981228 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:08 crc kubenswrapper[4946]: I1203 07:14:08.981247 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf9fm\" (UniqueName: \"kubernetes.io/projected/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-kube-api-access-lf9fm\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.411581 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" event={"ID":"c6109109-dfb1-4b8c-86c7-94d7fe00fed1","Type":"ContainerDied","Data":"22dcb657c0ebe3eba25472b4a4409735ccf6760c22153bf82e198f2f2cb2b0a7"} Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.411636 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22dcb657c0ebe3eba25472b4a4409735ccf6760c22153bf82e198f2f2cb2b0a7" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.411684 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xp9sr" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.490230 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data\") pod \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\" (UID: \"c6109109-dfb1-4b8c-86c7-94d7fe00fed1\") " Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.498561 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data" (OuterVolumeSpecName: "config-data") pod "c6109109-dfb1-4b8c-86c7-94d7fe00fed1" (UID: "c6109109-dfb1-4b8c-86c7-94d7fe00fed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.559592 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 07:14:09 crc kubenswrapper[4946]: E1203 07:14:09.560081 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6109109-dfb1-4b8c-86c7-94d7fe00fed1" containerName="nova-cell0-conductor-db-sync" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.560102 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6109109-dfb1-4b8c-86c7-94d7fe00fed1" containerName="nova-cell0-conductor-db-sync" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.560312 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6109109-dfb1-4b8c-86c7-94d7fe00fed1" containerName="nova-cell0-conductor-db-sync" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.561013 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.571327 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.591975 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6109109-dfb1-4b8c-86c7-94d7fe00fed1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.694551 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.694781 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.694835 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwtk2\" (UniqueName: \"kubernetes.io/projected/0e0b4114-0121-48cc-9439-bc5d350aeb44-kube-api-access-fwtk2\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.796648 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.797044 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwtk2\" (UniqueName: \"kubernetes.io/projected/0e0b4114-0121-48cc-9439-bc5d350aeb44-kube-api-access-fwtk2\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.797139 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.802439 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.803967 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.814616 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwtk2\" (UniqueName: \"kubernetes.io/projected/0e0b4114-0121-48cc-9439-bc5d350aeb44-kube-api-access-fwtk2\") pod \"nova-cell0-conductor-0\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:09 crc kubenswrapper[4946]: I1203 07:14:09.893730 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:10 crc kubenswrapper[4946]: I1203 07:14:10.375968 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 07:14:10 crc kubenswrapper[4946]: I1203 07:14:10.426510 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0e0b4114-0121-48cc-9439-bc5d350aeb44","Type":"ContainerStarted","Data":"2a7a80c3ff2375af3de6f56549364816f5acd049a27afae2840afd3d289e3190"} Dec 03 07:14:11 crc kubenswrapper[4946]: I1203 07:14:11.443492 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0e0b4114-0121-48cc-9439-bc5d350aeb44","Type":"ContainerStarted","Data":"d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217"} Dec 03 07:14:11 crc kubenswrapper[4946]: I1203 07:14:11.443998 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:11 crc kubenswrapper[4946]: I1203 07:14:11.470850 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.470814376 podStartE2EDuration="2.470814376s" podCreationTimestamp="2025-12-03 07:14:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:11.460314775 +0000 UTC m=+1444.257004934" watchObservedRunningTime="2025-12-03 07:14:11.470814376 +0000 UTC m=+1444.267504535" Dec 03 07:14:19 crc kubenswrapper[4946]: I1203 07:14:19.940719 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.411441 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-mdtz2"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.413092 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.416729 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.416806 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.424415 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mdtz2"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.496281 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.496331 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-scripts\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.496424 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-config-data\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.496466 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vncp8\" (UniqueName: \"kubernetes.io/projected/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-kube-api-access-vncp8\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.554094 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.555167 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.556813 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.582308 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.634438 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.634489 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfvfk\" (UniqueName: \"kubernetes.io/projected/e4962658-f138-4157-8ec0-a4d5d951711d-kube-api-access-kfvfk\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.634552 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.634948 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.634969 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-scripts\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.635493 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-config-data\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.635543 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vncp8\" (UniqueName: \"kubernetes.io/projected/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-kube-api-access-vncp8\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.644006 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-config-data\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.647110 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.650433 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-scripts\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.665960 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vncp8\" (UniqueName: \"kubernetes.io/projected/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-kube-api-access-vncp8\") pod \"nova-cell0-cell-mapping-mdtz2\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.737176 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.737225 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfvfk\" (UniqueName: \"kubernetes.io/projected/e4962658-f138-4157-8ec0-a4d5d951711d-kube-api-access-kfvfk\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.737262 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.740485 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.740609 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.741221 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.756309 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.757507 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.760417 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.769463 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfvfk\" (UniqueName: \"kubernetes.io/projected/e4962658-f138-4157-8ec0-a4d5d951711d-kube-api-access-kfvfk\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.803691 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.830428 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.832052 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.834222 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.841444 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5ptk\" (UniqueName: \"kubernetes.io/projected/7d1d5102-9075-4423-8735-41834caa451a-kube-api-access-q5ptk\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.841535 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.841559 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-config-data\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.841580 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d1d5102-9075-4423-8735-41834caa451a-logs\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.841630 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-config-data\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.841695 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.842547 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmg72\" (UniqueName: \"kubernetes.io/projected/05423f2d-1369-485b-ba07-bcda3b99fa83-kube-api-access-fmg72\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.879919 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.898027 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.928289 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.936306 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.937175 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.948599 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-config-data\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.949841 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d1d5102-9075-4423-8735-41834caa451a-logs\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.949984 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-config-data\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.950111 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.950253 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmg72\" (UniqueName: \"kubernetes.io/projected/05423f2d-1369-485b-ba07-bcda3b99fa83-kube-api-access-fmg72\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.950292 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5ptk\" (UniqueName: \"kubernetes.io/projected/7d1d5102-9075-4423-8735-41834caa451a-kube-api-access-q5ptk\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.950402 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.952438 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d1d5102-9075-4423-8735-41834caa451a-logs\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.957625 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-config-data\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.957670 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.976626 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-config-data\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:20 crc kubenswrapper[4946]: I1203 07:14:20.977267 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.006367 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmg72\" (UniqueName: \"kubernetes.io/projected/05423f2d-1369-485b-ba07-bcda3b99fa83-kube-api-access-fmg72\") pod \"nova-scheduler-0\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.007479 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5ptk\" (UniqueName: \"kubernetes.io/projected/7d1d5102-9075-4423-8735-41834caa451a-kube-api-access-q5ptk\") pod \"nova-api-0\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " pod="openstack/nova-api-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.091119 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.093860 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a607344a-aa2e-4456-aa1f-e686e6d19ed9-logs\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.093952 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-config-data\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.093978 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.094043 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfxgc\" (UniqueName: \"kubernetes.io/projected/a607344a-aa2e-4456-aa1f-e686e6d19ed9-kube-api-access-dfxgc\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.188812 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b9ff45c7-mpjl5"] Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.190371 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.197633 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfxgc\" (UniqueName: \"kubernetes.io/projected/a607344a-aa2e-4456-aa1f-e686e6d19ed9-kube-api-access-dfxgc\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.197721 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a607344a-aa2e-4456-aa1f-e686e6d19ed9-logs\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.197798 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-config-data\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.197821 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.199108 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a607344a-aa2e-4456-aa1f-e686e6d19ed9-logs\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.204596 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.208779 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-config-data\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.209228 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.216350 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b9ff45c7-mpjl5"] Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.239803 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.242473 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfxgc\" (UniqueName: \"kubernetes.io/projected/a607344a-aa2e-4456-aa1f-e686e6d19ed9-kube-api-access-dfxgc\") pod \"nova-metadata-0\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.263165 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.299397 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-swift-storage-0\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.299445 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-config\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.299487 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-nb\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.299578 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-sb\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.299624 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64wz8\" (UniqueName: \"kubernetes.io/projected/b48f8844-fcc2-4892-826b-47e71963274f-kube-api-access-64wz8\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.299659 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-svc\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.404940 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-swift-storage-0\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.405015 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-config\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.405056 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-nb\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.405120 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-sb\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.405144 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64wz8\" (UniqueName: \"kubernetes.io/projected/b48f8844-fcc2-4892-826b-47e71963274f-kube-api-access-64wz8\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.405176 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-svc\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.406036 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-swift-storage-0\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.406114 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-svc\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.406712 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-nb\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.409208 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-config\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.409710 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-sb\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.437165 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64wz8\" (UniqueName: \"kubernetes.io/projected/b48f8844-fcc2-4892-826b-47e71963274f-kube-api-access-64wz8\") pod \"dnsmasq-dns-b9ff45c7-mpjl5\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.495859 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mdtz2"] Dec 03 07:14:21 crc kubenswrapper[4946]: W1203 07:14:21.496361 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778a6a04_4fc8_46fc_8f83_9d3625eb93d8.slice/crio-2259ef4593dc51a8e625c19549230b30b2f134204a185ef54185b7838336cc4d WatchSource:0}: Error finding container 2259ef4593dc51a8e625c19549230b30b2f134204a185ef54185b7838336cc4d: Status 404 returned error can't find the container with id 2259ef4593dc51a8e625c19549230b30b2f134204a185ef54185b7838336cc4d Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.519844 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.580195 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mdtz2" event={"ID":"778a6a04-4fc8-46fc-8f83-9d3625eb93d8","Type":"ContainerStarted","Data":"2259ef4593dc51a8e625c19549230b30b2f134204a185ef54185b7838336cc4d"} Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.657112 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.794998 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.910821 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:21 crc kubenswrapper[4946]: W1203 07:14:21.926539 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda607344a_aa2e_4456_aa1f_e686e6d19ed9.slice/crio-318333f1af3c8b1341a5cf4988b51dfd749d9b99ea4c52b630bf70f2cca68c0e WatchSource:0}: Error finding container 318333f1af3c8b1341a5cf4988b51dfd749d9b99ea4c52b630bf70f2cca68c0e: Status 404 returned error can't find the container with id 318333f1af3c8b1341a5cf4988b51dfd749d9b99ea4c52b630bf70f2cca68c0e Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.949938 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.963897 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8vff8"] Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.965328 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.968878 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.969051 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 07:14:21 crc kubenswrapper[4946]: I1203 07:14:21.984289 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8vff8"] Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.086263 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b9ff45c7-mpjl5"] Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.125498 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56nhg\" (UniqueName: \"kubernetes.io/projected/70f4cb0d-e6fd-45dc-be44-40233bad9532-kube-api-access-56nhg\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.125638 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.126138 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-config-data\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.126320 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-scripts\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.228154 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-config-data\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.228813 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-scripts\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.228893 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56nhg\" (UniqueName: \"kubernetes.io/projected/70f4cb0d-e6fd-45dc-be44-40233bad9532-kube-api-access-56nhg\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.228959 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.233189 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.233451 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-scripts\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.235414 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-config-data\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.245515 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56nhg\" (UniqueName: \"kubernetes.io/projected/70f4cb0d-e6fd-45dc-be44-40233bad9532-kube-api-access-56nhg\") pod \"nova-cell1-conductor-db-sync-8vff8\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.291928 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.593151 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e4962658-f138-4157-8ec0-a4d5d951711d","Type":"ContainerStarted","Data":"d56766efee9d3aa1ec20fb1da046a1b9f36c4211f24bb0cdd980c79cea062768"} Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.598010 4946 generic.go:334] "Generic (PLEG): container finished" podID="b48f8844-fcc2-4892-826b-47e71963274f" containerID="76446d6826fb88bd3875eab86229dd6c5b12e905abcac321999b9ba5895aee85" exitCode=0 Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.598075 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" event={"ID":"b48f8844-fcc2-4892-826b-47e71963274f","Type":"ContainerDied","Data":"76446d6826fb88bd3875eab86229dd6c5b12e905abcac321999b9ba5895aee85"} Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.598102 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" event={"ID":"b48f8844-fcc2-4892-826b-47e71963274f","Type":"ContainerStarted","Data":"deb8b166cac0585b69cbb4f4de2b030fdfb02b9334af953c9901ed0e45071914"} Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.624994 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a607344a-aa2e-4456-aa1f-e686e6d19ed9","Type":"ContainerStarted","Data":"318333f1af3c8b1341a5cf4988b51dfd749d9b99ea4c52b630bf70f2cca68c0e"} Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.627641 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d1d5102-9075-4423-8735-41834caa451a","Type":"ContainerStarted","Data":"e4ccfa04a8d18a7c9b28c5c6632f6f4b8d7f9c64fb7f01cd1768cb4666a5cd0c"} Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.634541 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mdtz2" event={"ID":"778a6a04-4fc8-46fc-8f83-9d3625eb93d8","Type":"ContainerStarted","Data":"b4d0c049e4ebf0ba2b784151e0a28518ad34dc7f7cabb79dbe4c2f1ad0f9a64a"} Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.638783 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"05423f2d-1369-485b-ba07-bcda3b99fa83","Type":"ContainerStarted","Data":"a74941f3509359efe45275672e54a1dab0b82e507460404a7bb257608c495b61"} Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.654565 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-mdtz2" podStartSLOduration=2.654542376 podStartE2EDuration="2.654542376s" podCreationTimestamp="2025-12-03 07:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:22.649349238 +0000 UTC m=+1455.446039347" watchObservedRunningTime="2025-12-03 07:14:22.654542376 +0000 UTC m=+1455.451232485" Dec 03 07:14:22 crc kubenswrapper[4946]: I1203 07:14:22.784772 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8vff8"] Dec 03 07:14:23 crc kubenswrapper[4946]: I1203 07:14:23.660656 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" event={"ID":"b48f8844-fcc2-4892-826b-47e71963274f","Type":"ContainerStarted","Data":"39d9643b1a86ee3d6d3b7e80844ec0028ee0f00e06200359e250cf87ad5cf169"} Dec 03 07:14:23 crc kubenswrapper[4946]: I1203 07:14:23.661061 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:23 crc kubenswrapper[4946]: I1203 07:14:23.661903 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8vff8" event={"ID":"70f4cb0d-e6fd-45dc-be44-40233bad9532","Type":"ContainerStarted","Data":"2053513fb2cd98af806d2471e5ac121d3b4f53386b03f6ab961cfeb726f810cd"} Dec 03 07:14:23 crc kubenswrapper[4946]: I1203 07:14:23.661924 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8vff8" event={"ID":"70f4cb0d-e6fd-45dc-be44-40233bad9532","Type":"ContainerStarted","Data":"11955bb5e63e9075afd4f5de39bdb211d2ad8885276e8d2af87780064f03a5a6"} Dec 03 07:14:23 crc kubenswrapper[4946]: I1203 07:14:23.704223 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" podStartSLOduration=3.704201735 podStartE2EDuration="3.704201735s" podCreationTimestamp="2025-12-03 07:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:23.686119062 +0000 UTC m=+1456.482809171" watchObservedRunningTime="2025-12-03 07:14:23.704201735 +0000 UTC m=+1456.500891874" Dec 03 07:14:23 crc kubenswrapper[4946]: I1203 07:14:23.730067 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8vff8" podStartSLOduration=2.7300457160000002 podStartE2EDuration="2.730045716s" podCreationTimestamp="2025-12-03 07:14:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:23.699413117 +0000 UTC m=+1456.496103256" watchObservedRunningTime="2025-12-03 07:14:23.730045716 +0000 UTC m=+1456.526735825" Dec 03 07:14:24 crc kubenswrapper[4946]: I1203 07:14:24.735315 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:24 crc kubenswrapper[4946]: I1203 07:14:24.748939 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:25 crc kubenswrapper[4946]: I1203 07:14:25.697631 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e4962658-f138-4157-8ec0-a4d5d951711d","Type":"ContainerStarted","Data":"6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df"} Dec 03 07:14:25 crc kubenswrapper[4946]: I1203 07:14:25.697762 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e4962658-f138-4157-8ec0-a4d5d951711d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df" gracePeriod=30 Dec 03 07:14:25 crc kubenswrapper[4946]: I1203 07:14:25.701778 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a607344a-aa2e-4456-aa1f-e686e6d19ed9","Type":"ContainerStarted","Data":"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911"} Dec 03 07:14:25 crc kubenswrapper[4946]: I1203 07:14:25.708728 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d1d5102-9075-4423-8735-41834caa451a","Type":"ContainerStarted","Data":"7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f"} Dec 03 07:14:25 crc kubenswrapper[4946]: I1203 07:14:25.711427 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"05423f2d-1369-485b-ba07-bcda3b99fa83","Type":"ContainerStarted","Data":"f8831af5238f203cf08c7dcc6eddea95137d71a86314174f056b6982d075405b"} Dec 03 07:14:25 crc kubenswrapper[4946]: I1203 07:14:25.724626 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.185910958 podStartE2EDuration="5.724561923s" podCreationTimestamp="2025-12-03 07:14:20 +0000 UTC" firstStartedPulling="2025-12-03 07:14:21.680600031 +0000 UTC m=+1454.477290140" lastFinishedPulling="2025-12-03 07:14:25.219250996 +0000 UTC m=+1458.015941105" observedRunningTime="2025-12-03 07:14:25.714442353 +0000 UTC m=+1458.511132482" watchObservedRunningTime="2025-12-03 07:14:25.724561923 +0000 UTC m=+1458.521252042" Dec 03 07:14:25 crc kubenswrapper[4946]: I1203 07:14:25.737460 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.404032064 podStartE2EDuration="5.737443767s" podCreationTimestamp="2025-12-03 07:14:20 +0000 UTC" firstStartedPulling="2025-12-03 07:14:21.889638794 +0000 UTC m=+1454.686328893" lastFinishedPulling="2025-12-03 07:14:25.223050447 +0000 UTC m=+1458.019740596" observedRunningTime="2025-12-03 07:14:25.730489001 +0000 UTC m=+1458.527179110" watchObservedRunningTime="2025-12-03 07:14:25.737443767 +0000 UTC m=+1458.534133876" Dec 03 07:14:25 crc kubenswrapper[4946]: I1203 07:14:25.938298 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:26 crc kubenswrapper[4946]: I1203 07:14:26.210227 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 07:14:26 crc kubenswrapper[4946]: I1203 07:14:26.697178 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 07:14:26 crc kubenswrapper[4946]: I1203 07:14:26.723994 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d1d5102-9075-4423-8735-41834caa451a","Type":"ContainerStarted","Data":"91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551"} Dec 03 07:14:26 crc kubenswrapper[4946]: I1203 07:14:26.735855 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerName="nova-metadata-log" containerID="cri-o://52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911" gracePeriod=30 Dec 03 07:14:26 crc kubenswrapper[4946]: I1203 07:14:26.736224 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerName="nova-metadata-metadata" containerID="cri-o://fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19" gracePeriod=30 Dec 03 07:14:26 crc kubenswrapper[4946]: I1203 07:14:26.736264 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a607344a-aa2e-4456-aa1f-e686e6d19ed9","Type":"ContainerStarted","Data":"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19"} Dec 03 07:14:26 crc kubenswrapper[4946]: I1203 07:14:26.783480 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.378988529 podStartE2EDuration="6.783462479s" podCreationTimestamp="2025-12-03 07:14:20 +0000 UTC" firstStartedPulling="2025-12-03 07:14:21.81606033 +0000 UTC m=+1454.612750439" lastFinishedPulling="2025-12-03 07:14:25.22053425 +0000 UTC m=+1458.017224389" observedRunningTime="2025-12-03 07:14:26.755365308 +0000 UTC m=+1459.552055417" watchObservedRunningTime="2025-12-03 07:14:26.783462479 +0000 UTC m=+1459.580152588" Dec 03 07:14:26 crc kubenswrapper[4946]: I1203 07:14:26.816318 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.526504478 podStartE2EDuration="6.816301186s" podCreationTimestamp="2025-12-03 07:14:20 +0000 UTC" firstStartedPulling="2025-12-03 07:14:21.932226162 +0000 UTC m=+1454.728916271" lastFinishedPulling="2025-12-03 07:14:25.22202283 +0000 UTC m=+1458.018712979" observedRunningTime="2025-12-03 07:14:26.799197829 +0000 UTC m=+1459.595887938" watchObservedRunningTime="2025-12-03 07:14:26.816301186 +0000 UTC m=+1459.612991295" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.297127 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.441401 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a607344a-aa2e-4456-aa1f-e686e6d19ed9-logs\") pod \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.441450 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-config-data\") pod \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.441520 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-combined-ca-bundle\") pod \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.441672 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfxgc\" (UniqueName: \"kubernetes.io/projected/a607344a-aa2e-4456-aa1f-e686e6d19ed9-kube-api-access-dfxgc\") pod \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\" (UID: \"a607344a-aa2e-4456-aa1f-e686e6d19ed9\") " Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.443104 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a607344a-aa2e-4456-aa1f-e686e6d19ed9-logs" (OuterVolumeSpecName: "logs") pod "a607344a-aa2e-4456-aa1f-e686e6d19ed9" (UID: "a607344a-aa2e-4456-aa1f-e686e6d19ed9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.447085 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a607344a-aa2e-4456-aa1f-e686e6d19ed9-kube-api-access-dfxgc" (OuterVolumeSpecName: "kube-api-access-dfxgc") pod "a607344a-aa2e-4456-aa1f-e686e6d19ed9" (UID: "a607344a-aa2e-4456-aa1f-e686e6d19ed9"). InnerVolumeSpecName "kube-api-access-dfxgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.473258 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a607344a-aa2e-4456-aa1f-e686e6d19ed9" (UID: "a607344a-aa2e-4456-aa1f-e686e6d19ed9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.483903 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-config-data" (OuterVolumeSpecName: "config-data") pod "a607344a-aa2e-4456-aa1f-e686e6d19ed9" (UID: "a607344a-aa2e-4456-aa1f-e686e6d19ed9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.543572 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfxgc\" (UniqueName: \"kubernetes.io/projected/a607344a-aa2e-4456-aa1f-e686e6d19ed9-kube-api-access-dfxgc\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.543602 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a607344a-aa2e-4456-aa1f-e686e6d19ed9-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.543616 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.543651 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a607344a-aa2e-4456-aa1f-e686e6d19ed9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.747834 4946 generic.go:334] "Generic (PLEG): container finished" podID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerID="fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19" exitCode=0 Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.747864 4946 generic.go:334] "Generic (PLEG): container finished" podID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerID="52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911" exitCode=143 Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.747880 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.747924 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a607344a-aa2e-4456-aa1f-e686e6d19ed9","Type":"ContainerDied","Data":"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19"} Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.747949 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a607344a-aa2e-4456-aa1f-e686e6d19ed9","Type":"ContainerDied","Data":"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911"} Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.747961 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a607344a-aa2e-4456-aa1f-e686e6d19ed9","Type":"ContainerDied","Data":"318333f1af3c8b1341a5cf4988b51dfd749d9b99ea4c52b630bf70f2cca68c0e"} Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.747978 4946 scope.go:117] "RemoveContainer" containerID="fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.770111 4946 scope.go:117] "RemoveContainer" containerID="52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.781132 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.789071 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.797024 4946 scope.go:117] "RemoveContainer" containerID="fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19" Dec 03 07:14:27 crc kubenswrapper[4946]: E1203 07:14:27.797716 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19\": container with ID starting with fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19 not found: ID does not exist" containerID="fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.797820 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19"} err="failed to get container status \"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19\": rpc error: code = NotFound desc = could not find container \"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19\": container with ID starting with fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19 not found: ID does not exist" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.797887 4946 scope.go:117] "RemoveContainer" containerID="52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911" Dec 03 07:14:27 crc kubenswrapper[4946]: E1203 07:14:27.798424 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911\": container with ID starting with 52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911 not found: ID does not exist" containerID="52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.798456 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911"} err="failed to get container status \"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911\": rpc error: code = NotFound desc = could not find container \"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911\": container with ID starting with 52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911 not found: ID does not exist" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.798477 4946 scope.go:117] "RemoveContainer" containerID="fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.798784 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19"} err="failed to get container status \"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19\": rpc error: code = NotFound desc = could not find container \"fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19\": container with ID starting with fafe876476969539694f6d4e4fdfe4b0569bbd47d7baaa2836b932e1ed30ab19 not found: ID does not exist" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.798803 4946 scope.go:117] "RemoveContainer" containerID="52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.799103 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911"} err="failed to get container status \"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911\": rpc error: code = NotFound desc = could not find container \"52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911\": container with ID starting with 52175c8ca0a5b8b5edbefb57cd3842b4ba74b5731d63630618680b4e906e9911 not found: ID does not exist" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.809149 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:27 crc kubenswrapper[4946]: E1203 07:14:27.809790 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerName="nova-metadata-metadata" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.809823 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerName="nova-metadata-metadata" Dec 03 07:14:27 crc kubenswrapper[4946]: E1203 07:14:27.809854 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerName="nova-metadata-log" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.809866 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerName="nova-metadata-log" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.812378 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerName="nova-metadata-log" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.812442 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" containerName="nova-metadata-metadata" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.814124 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.818865 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.819513 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.823218 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.951830 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.952175 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-config-data\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.952226 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.952269 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvdtx\" (UniqueName: \"kubernetes.io/projected/26da55ad-93ff-423c-b0c8-0183983e19f7-kube-api-access-hvdtx\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:27 crc kubenswrapper[4946]: I1203 07:14:27.952393 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26da55ad-93ff-423c-b0c8-0183983e19f7-logs\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.054111 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26da55ad-93ff-423c-b0c8-0183983e19f7-logs\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.054189 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.054220 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-config-data\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.054266 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.054309 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvdtx\" (UniqueName: \"kubernetes.io/projected/26da55ad-93ff-423c-b0c8-0183983e19f7-kube-api-access-hvdtx\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.055173 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26da55ad-93ff-423c-b0c8-0183983e19f7-logs\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.060312 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.060667 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-config-data\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.060891 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.074076 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvdtx\" (UniqueName: \"kubernetes.io/projected/26da55ad-93ff-423c-b0c8-0183983e19f7-kube-api-access-hvdtx\") pod \"nova-metadata-0\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.173518 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:28 crc kubenswrapper[4946]: W1203 07:14:28.688648 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26da55ad_93ff_423c_b0c8_0183983e19f7.slice/crio-27a9581cef15fa03aab9b766259882d610157e31b735758545809284b405ad63 WatchSource:0}: Error finding container 27a9581cef15fa03aab9b766259882d610157e31b735758545809284b405ad63: Status 404 returned error can't find the container with id 27a9581cef15fa03aab9b766259882d610157e31b735758545809284b405ad63 Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.700342 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:28 crc kubenswrapper[4946]: I1203 07:14:28.761316 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26da55ad-93ff-423c-b0c8-0183983e19f7","Type":"ContainerStarted","Data":"27a9581cef15fa03aab9b766259882d610157e31b735758545809284b405ad63"} Dec 03 07:14:29 crc kubenswrapper[4946]: I1203 07:14:29.603226 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a607344a-aa2e-4456-aa1f-e686e6d19ed9" path="/var/lib/kubelet/pods/a607344a-aa2e-4456-aa1f-e686e6d19ed9/volumes" Dec 03 07:14:29 crc kubenswrapper[4946]: I1203 07:14:29.780394 4946 generic.go:334] "Generic (PLEG): container finished" podID="778a6a04-4fc8-46fc-8f83-9d3625eb93d8" containerID="b4d0c049e4ebf0ba2b784151e0a28518ad34dc7f7cabb79dbe4c2f1ad0f9a64a" exitCode=0 Dec 03 07:14:29 crc kubenswrapper[4946]: I1203 07:14:29.780556 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mdtz2" event={"ID":"778a6a04-4fc8-46fc-8f83-9d3625eb93d8","Type":"ContainerDied","Data":"b4d0c049e4ebf0ba2b784151e0a28518ad34dc7f7cabb79dbe4c2f1ad0f9a64a"} Dec 03 07:14:29 crc kubenswrapper[4946]: I1203 07:14:29.783262 4946 generic.go:334] "Generic (PLEG): container finished" podID="70f4cb0d-e6fd-45dc-be44-40233bad9532" containerID="2053513fb2cd98af806d2471e5ac121d3b4f53386b03f6ab961cfeb726f810cd" exitCode=0 Dec 03 07:14:29 crc kubenswrapper[4946]: I1203 07:14:29.783372 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8vff8" event={"ID":"70f4cb0d-e6fd-45dc-be44-40233bad9532","Type":"ContainerDied","Data":"2053513fb2cd98af806d2471e5ac121d3b4f53386b03f6ab961cfeb726f810cd"} Dec 03 07:14:29 crc kubenswrapper[4946]: I1203 07:14:29.786577 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26da55ad-93ff-423c-b0c8-0183983e19f7","Type":"ContainerStarted","Data":"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b"} Dec 03 07:14:29 crc kubenswrapper[4946]: I1203 07:14:29.786724 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26da55ad-93ff-423c-b0c8-0183983e19f7","Type":"ContainerStarted","Data":"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2"} Dec 03 07:14:29 crc kubenswrapper[4946]: I1203 07:14:29.843435 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.8434152360000002 podStartE2EDuration="2.843415236s" podCreationTimestamp="2025-12-03 07:14:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:29.842083851 +0000 UTC m=+1462.638773950" watchObservedRunningTime="2025-12-03 07:14:29.843415236 +0000 UTC m=+1462.640105345" Dec 03 07:14:30 crc kubenswrapper[4946]: I1203 07:14:30.695205 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:14:30 crc kubenswrapper[4946]: I1203 07:14:30.696553 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e0916ede-42cd-4da1-86f0-db41c56c64ec" containerName="kube-state-metrics" containerID="cri-o://41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5" gracePeriod=30 Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.173031 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.209587 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.242222 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.242282 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.250891 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.313385 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-scripts\") pod \"70f4cb0d-e6fd-45dc-be44-40233bad9532\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.313470 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-combined-ca-bundle\") pod \"70f4cb0d-e6fd-45dc-be44-40233bad9532\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.313506 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56nhg\" (UniqueName: \"kubernetes.io/projected/70f4cb0d-e6fd-45dc-be44-40233bad9532-kube-api-access-56nhg\") pod \"70f4cb0d-e6fd-45dc-be44-40233bad9532\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.313570 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-config-data\") pod \"70f4cb0d-e6fd-45dc-be44-40233bad9532\" (UID: \"70f4cb0d-e6fd-45dc-be44-40233bad9532\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.319146 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f4cb0d-e6fd-45dc-be44-40233bad9532-kube-api-access-56nhg" (OuterVolumeSpecName: "kube-api-access-56nhg") pod "70f4cb0d-e6fd-45dc-be44-40233bad9532" (UID: "70f4cb0d-e6fd-45dc-be44-40233bad9532"). InnerVolumeSpecName "kube-api-access-56nhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.319616 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-scripts" (OuterVolumeSpecName: "scripts") pod "70f4cb0d-e6fd-45dc-be44-40233bad9532" (UID: "70f4cb0d-e6fd-45dc-be44-40233bad9532"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.341593 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70f4cb0d-e6fd-45dc-be44-40233bad9532" (UID: "70f4cb0d-e6fd-45dc-be44-40233bad9532"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.346816 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-config-data" (OuterVolumeSpecName: "config-data") pod "70f4cb0d-e6fd-45dc-be44-40233bad9532" (UID: "70f4cb0d-e6fd-45dc-be44-40233bad9532"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.363890 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.369882 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.427385 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.427415 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.427426 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56nhg\" (UniqueName: \"kubernetes.io/projected/70f4cb0d-e6fd-45dc-be44-40233bad9532-kube-api-access-56nhg\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.427436 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f4cb0d-e6fd-45dc-be44-40233bad9532-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.522004 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.528670 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-combined-ca-bundle\") pod \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.528923 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-scripts\") pod \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.529103 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vncp8\" (UniqueName: \"kubernetes.io/projected/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-kube-api-access-vncp8\") pod \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.529213 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-config-data\") pod \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\" (UID: \"778a6a04-4fc8-46fc-8f83-9d3625eb93d8\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.529432 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxjd7\" (UniqueName: \"kubernetes.io/projected/e0916ede-42cd-4da1-86f0-db41c56c64ec-kube-api-access-cxjd7\") pod \"e0916ede-42cd-4da1-86f0-db41c56c64ec\" (UID: \"e0916ede-42cd-4da1-86f0-db41c56c64ec\") " Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.534830 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-kube-api-access-vncp8" (OuterVolumeSpecName: "kube-api-access-vncp8") pod "778a6a04-4fc8-46fc-8f83-9d3625eb93d8" (UID: "778a6a04-4fc8-46fc-8f83-9d3625eb93d8"). InnerVolumeSpecName "kube-api-access-vncp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.536411 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-scripts" (OuterVolumeSpecName: "scripts") pod "778a6a04-4fc8-46fc-8f83-9d3625eb93d8" (UID: "778a6a04-4fc8-46fc-8f83-9d3625eb93d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.541360 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0916ede-42cd-4da1-86f0-db41c56c64ec-kube-api-access-cxjd7" (OuterVolumeSpecName: "kube-api-access-cxjd7") pod "e0916ede-42cd-4da1-86f0-db41c56c64ec" (UID: "e0916ede-42cd-4da1-86f0-db41c56c64ec"). InnerVolumeSpecName "kube-api-access-cxjd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.559528 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-config-data" (OuterVolumeSpecName: "config-data") pod "778a6a04-4fc8-46fc-8f83-9d3625eb93d8" (UID: "778a6a04-4fc8-46fc-8f83-9d3625eb93d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.601962 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "778a6a04-4fc8-46fc-8f83-9d3625eb93d8" (UID: "778a6a04-4fc8-46fc-8f83-9d3625eb93d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.622038 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-894d58c65-lxppn"] Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.623918 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-894d58c65-lxppn" podUID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" containerName="dnsmasq-dns" containerID="cri-o://71e308f35c992bfb97c644a69f108037c01605c3e14403a29286ed29fc5a3878" gracePeriod=10 Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.634578 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxjd7\" (UniqueName: \"kubernetes.io/projected/e0916ede-42cd-4da1-86f0-db41c56c64ec-kube-api-access-cxjd7\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.634717 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.634793 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.634865 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vncp8\" (UniqueName: \"kubernetes.io/projected/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-kube-api-access-vncp8\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.634920 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778a6a04-4fc8-46fc-8f83-9d3625eb93d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.809991 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mdtz2" event={"ID":"778a6a04-4fc8-46fc-8f83-9d3625eb93d8","Type":"ContainerDied","Data":"2259ef4593dc51a8e625c19549230b30b2f134204a185ef54185b7838336cc4d"} Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.810041 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2259ef4593dc51a8e625c19549230b30b2f134204a185ef54185b7838336cc4d" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.810123 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mdtz2" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.814988 4946 generic.go:334] "Generic (PLEG): container finished" podID="e0916ede-42cd-4da1-86f0-db41c56c64ec" containerID="41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5" exitCode=2 Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.815041 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e0916ede-42cd-4da1-86f0-db41c56c64ec","Type":"ContainerDied","Data":"41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5"} Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.815062 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e0916ede-42cd-4da1-86f0-db41c56c64ec","Type":"ContainerDied","Data":"cbd535d9e8ea1233707a0e8fcb2b67b8c22be494735759541aa75cfa4974e045"} Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.815077 4946 scope.go:117] "RemoveContainer" containerID="41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.815179 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.819646 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8vff8" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.819663 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8vff8" event={"ID":"70f4cb0d-e6fd-45dc-be44-40233bad9532","Type":"ContainerDied","Data":"11955bb5e63e9075afd4f5de39bdb211d2ad8885276e8d2af87780064f03a5a6"} Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.819709 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11955bb5e63e9075afd4f5de39bdb211d2ad8885276e8d2af87780064f03a5a6" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.828293 4946 generic.go:334] "Generic (PLEG): container finished" podID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" containerID="71e308f35c992bfb97c644a69f108037c01605c3e14403a29286ed29fc5a3878" exitCode=0 Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.828330 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-894d58c65-lxppn" event={"ID":"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc","Type":"ContainerDied","Data":"71e308f35c992bfb97c644a69f108037c01605c3e14403a29286ed29fc5a3878"} Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.877767 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.903541 4946 scope.go:117] "RemoveContainer" containerID="41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5" Dec 03 07:14:31 crc kubenswrapper[4946]: E1203 07:14:31.905414 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5\": container with ID starting with 41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5 not found: ID does not exist" containerID="41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.905456 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5"} err="failed to get container status \"41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5\": rpc error: code = NotFound desc = could not find container \"41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5\": container with ID starting with 41fb44b5b6fcd9fe5cc61eef4526233afb1bcdf44f7569d397764fcab4aed7a5 not found: ID does not exist" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.909413 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 07:14:31 crc kubenswrapper[4946]: E1203 07:14:31.909896 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0916ede-42cd-4da1-86f0-db41c56c64ec" containerName="kube-state-metrics" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.909912 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0916ede-42cd-4da1-86f0-db41c56c64ec" containerName="kube-state-metrics" Dec 03 07:14:31 crc kubenswrapper[4946]: E1203 07:14:31.909933 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778a6a04-4fc8-46fc-8f83-9d3625eb93d8" containerName="nova-manage" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.909939 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="778a6a04-4fc8-46fc-8f83-9d3625eb93d8" containerName="nova-manage" Dec 03 07:14:31 crc kubenswrapper[4946]: E1203 07:14:31.909961 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f4cb0d-e6fd-45dc-be44-40233bad9532" containerName="nova-cell1-conductor-db-sync" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.909967 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f4cb0d-e6fd-45dc-be44-40233bad9532" containerName="nova-cell1-conductor-db-sync" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.910147 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="778a6a04-4fc8-46fc-8f83-9d3625eb93d8" containerName="nova-manage" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.910170 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f4cb0d-e6fd-45dc-be44-40233bad9532" containerName="nova-cell1-conductor-db-sync" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.910192 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0916ede-42cd-4da1-86f0-db41c56c64ec" containerName="kube-state-metrics" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.911544 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.921628 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.934184 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.964494 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:14:31 crc kubenswrapper[4946]: I1203 07:14:31.975010 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.013280 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.014644 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.017275 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.017408 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.034601 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.041444 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.041562 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.041609 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8drq\" (UniqueName: \"kubernetes.io/projected/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-kube-api-access-k8drq\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.045984 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.046222 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-log" containerID="cri-o://7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f" gracePeriod=30 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.046661 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-api" containerID="cri-o://91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551" gracePeriod=30 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.055962 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": EOF" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.058470 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": EOF" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.064569 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.064793 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerName="nova-metadata-log" containerID="cri-o://91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2" gracePeriod=30 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.064915 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerName="nova-metadata-metadata" containerID="cri-o://4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b" gracePeriod=30 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.144031 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.144181 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.144255 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.144313 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmxsp\" (UniqueName: \"kubernetes.io/projected/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-api-access-fmxsp\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.144376 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.144427 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8drq\" (UniqueName: \"kubernetes.io/projected/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-kube-api-access-k8drq\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.144510 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.148712 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.149172 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.162706 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8drq\" (UniqueName: \"kubernetes.io/projected/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-kube-api-access-k8drq\") pod \"nova-cell1-conductor-0\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.234634 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.245776 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.245820 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmxsp\" (UniqueName: \"kubernetes.io/projected/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-api-access-fmxsp\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.245892 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.245939 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.249892 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.250570 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.251169 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.269412 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmxsp\" (UniqueName: \"kubernetes.io/projected/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-api-access-fmxsp\") pod \"kube-state-metrics-0\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.341517 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.468791 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.652925 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.676223 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764528 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-sb\") pod \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764572 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-config-data\") pod \"26da55ad-93ff-423c-b0c8-0183983e19f7\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764612 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-nb\") pod \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764681 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvdtx\" (UniqueName: \"kubernetes.io/projected/26da55ad-93ff-423c-b0c8-0183983e19f7-kube-api-access-hvdtx\") pod \"26da55ad-93ff-423c-b0c8-0183983e19f7\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764759 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-combined-ca-bundle\") pod \"26da55ad-93ff-423c-b0c8-0183983e19f7\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764779 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26da55ad-93ff-423c-b0c8-0183983e19f7-logs\") pod \"26da55ad-93ff-423c-b0c8-0183983e19f7\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764798 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-config\") pod \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764891 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-svc\") pod \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764913 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dswlz\" (UniqueName: \"kubernetes.io/projected/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-kube-api-access-dswlz\") pod \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764941 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-swift-storage-0\") pod \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\" (UID: \"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.764965 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-nova-metadata-tls-certs\") pod \"26da55ad-93ff-423c-b0c8-0183983e19f7\" (UID: \"26da55ad-93ff-423c-b0c8-0183983e19f7\") " Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.767319 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26da55ad-93ff-423c-b0c8-0183983e19f7-logs" (OuterVolumeSpecName: "logs") pod "26da55ad-93ff-423c-b0c8-0183983e19f7" (UID: "26da55ad-93ff-423c-b0c8-0183983e19f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.770304 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-kube-api-access-dswlz" (OuterVolumeSpecName: "kube-api-access-dswlz") pod "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" (UID: "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc"). InnerVolumeSpecName "kube-api-access-dswlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.773904 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26da55ad-93ff-423c-b0c8-0183983e19f7-kube-api-access-hvdtx" (OuterVolumeSpecName: "kube-api-access-hvdtx") pod "26da55ad-93ff-423c-b0c8-0183983e19f7" (UID: "26da55ad-93ff-423c-b0c8-0183983e19f7"). InnerVolumeSpecName "kube-api-access-hvdtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.812445 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26da55ad-93ff-423c-b0c8-0183983e19f7" (UID: "26da55ad-93ff-423c-b0c8-0183983e19f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.842885 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-config-data" (OuterVolumeSpecName: "config-data") pod "26da55ad-93ff-423c-b0c8-0183983e19f7" (UID: "26da55ad-93ff-423c-b0c8-0183983e19f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.849828 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.850084 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="ceilometer-central-agent" containerID="cri-o://e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3" gracePeriod=30 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.850214 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="proxy-httpd" containerID="cri-o://3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c" gracePeriod=30 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.850249 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="sg-core" containerID="cri-o://0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85" gracePeriod=30 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.850297 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="ceilometer-notification-agent" containerID="cri-o://d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5" gracePeriod=30 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.862634 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-config" (OuterVolumeSpecName: "config") pod "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" (UID: "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867417 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867439 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26da55ad-93ff-423c-b0c8-0183983e19f7-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867449 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867460 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dswlz\" (UniqueName: \"kubernetes.io/projected/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-kube-api-access-dswlz\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867473 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867481 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvdtx\" (UniqueName: \"kubernetes.io/projected/26da55ad-93ff-423c-b0c8-0183983e19f7-kube-api-access-hvdtx\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867575 4946 generic.go:334] "Generic (PLEG): container finished" podID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerID="4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b" exitCode=0 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867591 4946 generic.go:334] "Generic (PLEG): container finished" podID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerID="91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2" exitCode=143 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867629 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26da55ad-93ff-423c-b0c8-0183983e19f7","Type":"ContainerDied","Data":"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b"} Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867655 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26da55ad-93ff-423c-b0c8-0183983e19f7","Type":"ContainerDied","Data":"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2"} Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867666 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26da55ad-93ff-423c-b0c8-0183983e19f7","Type":"ContainerDied","Data":"27a9581cef15fa03aab9b766259882d610157e31b735758545809284b405ad63"} Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.867681 4946 scope.go:117] "RemoveContainer" containerID="4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.868028 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.870220 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" (UID: "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.877683 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-894d58c65-lxppn" event={"ID":"3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc","Type":"ContainerDied","Data":"8cded015591d245505102a682dce67a1a4a288f981dcd3ac3800136d8ad55a59"} Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.877956 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-894d58c65-lxppn" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.893252 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.898301 4946 generic.go:334] "Generic (PLEG): container finished" podID="7d1d5102-9075-4423-8735-41834caa451a" containerID="7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f" exitCode=143 Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.898519 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d1d5102-9075-4423-8735-41834caa451a","Type":"ContainerDied","Data":"7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f"} Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.899832 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" (UID: "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.906279 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" (UID: "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.912390 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" (UID: "3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.914778 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "26da55ad-93ff-423c-b0c8-0183983e19f7" (UID: "26da55ad-93ff-423c-b0c8-0183983e19f7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:32 crc kubenswrapper[4946]: W1203 07:14:32.917056 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10d9fdfc_0d03_4ca5_ba0f_2f8c408b55de.slice/crio-e140144ec0a450c4367b9cfe0053b6718e433805a554f25af15235aaeb9f696d WatchSource:0}: Error finding container e140144ec0a450c4367b9cfe0053b6718e433805a554f25af15235aaeb9f696d: Status 404 returned error can't find the container with id e140144ec0a450c4367b9cfe0053b6718e433805a554f25af15235aaeb9f696d Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.969065 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.969097 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.969107 4946 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26da55ad-93ff-423c-b0c8-0183983e19f7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.969115 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:32 crc kubenswrapper[4946]: I1203 07:14:32.969124 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:33 crc kubenswrapper[4946]: W1203 07:14:33.043323 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf69ee496_d325_40e0_81ff_0fb3625a775a.slice/crio-961d002cc0d977266c82641602e07919bb97e9be177176ffddc7cd5b172c7e24 WatchSource:0}: Error finding container 961d002cc0d977266c82641602e07919bb97e9be177176ffddc7cd5b172c7e24: Status 404 returned error can't find the container with id 961d002cc0d977266c82641602e07919bb97e9be177176ffddc7cd5b172c7e24 Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.045484 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.115271 4946 scope.go:117] "RemoveContainer" containerID="91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.140121 4946 scope.go:117] "RemoveContainer" containerID="4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b" Dec 03 07:14:33 crc kubenswrapper[4946]: E1203 07:14:33.140689 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b\": container with ID starting with 4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b not found: ID does not exist" containerID="4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.140730 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b"} err="failed to get container status \"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b\": rpc error: code = NotFound desc = could not find container \"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b\": container with ID starting with 4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b not found: ID does not exist" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.140796 4946 scope.go:117] "RemoveContainer" containerID="91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2" Dec 03 07:14:33 crc kubenswrapper[4946]: E1203 07:14:33.141084 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2\": container with ID starting with 91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2 not found: ID does not exist" containerID="91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.141122 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2"} err="failed to get container status \"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2\": rpc error: code = NotFound desc = could not find container \"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2\": container with ID starting with 91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2 not found: ID does not exist" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.141142 4946 scope.go:117] "RemoveContainer" containerID="4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.141425 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b"} err="failed to get container status \"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b\": rpc error: code = NotFound desc = could not find container \"4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b\": container with ID starting with 4135a6f171e1b8e92f6b8cc9006cd0a5ddb93581c478f77dea8252c65c5a414b not found: ID does not exist" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.141446 4946 scope.go:117] "RemoveContainer" containerID="91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.141665 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2"} err="failed to get container status \"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2\": rpc error: code = NotFound desc = could not find container \"91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2\": container with ID starting with 91989455578b98edd1579942255e6d3bba43a60618b8f1e642323d89fc0b22a2 not found: ID does not exist" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.141686 4946 scope.go:117] "RemoveContainer" containerID="71e308f35c992bfb97c644a69f108037c01605c3e14403a29286ed29fc5a3878" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.184808 4946 scope.go:117] "RemoveContainer" containerID="4f80294365bb9ae4ed7c0a9ed42dbe052d5bad38103f1a6990bc0abf4581181f" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.226584 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.246368 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.263782 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-894d58c65-lxppn"] Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.314814 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-894d58c65-lxppn"] Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.327168 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:33 crc kubenswrapper[4946]: E1203 07:14:33.328101 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerName="nova-metadata-log" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.328118 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerName="nova-metadata-log" Dec 03 07:14:33 crc kubenswrapper[4946]: E1203 07:14:33.328128 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" containerName="init" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.328134 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" containerName="init" Dec 03 07:14:33 crc kubenswrapper[4946]: E1203 07:14:33.328193 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" containerName="dnsmasq-dns" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.328200 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" containerName="dnsmasq-dns" Dec 03 07:14:33 crc kubenswrapper[4946]: E1203 07:14:33.328231 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerName="nova-metadata-metadata" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.328238 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerName="nova-metadata-metadata" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.328650 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" containerName="dnsmasq-dns" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.328723 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerName="nova-metadata-metadata" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.328765 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" containerName="nova-metadata-log" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.330979 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.335944 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.348783 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.355957 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.507334 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlng5\" (UniqueName: \"kubernetes.io/projected/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-kube-api-access-qlng5\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.507426 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-config-data\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.507494 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-logs\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.507706 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.507774 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.606453 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26da55ad-93ff-423c-b0c8-0183983e19f7" path="/var/lib/kubelet/pods/26da55ad-93ff-423c-b0c8-0183983e19f7/volumes" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.607073 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc" path="/var/lib/kubelet/pods/3ef1e8cc-4747-4996-8e5c-5c3f31f1b5fc/volumes" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.607583 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0916ede-42cd-4da1-86f0-db41c56c64ec" path="/var/lib/kubelet/pods/e0916ede-42cd-4da1-86f0-db41c56c64ec/volumes" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.609146 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-config-data\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.609171 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-logs\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.609201 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.609223 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.609305 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlng5\" (UniqueName: \"kubernetes.io/projected/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-kube-api-access-qlng5\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.609990 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-logs\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.612915 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.613402 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-config-data\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.613430 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.626209 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlng5\" (UniqueName: \"kubernetes.io/projected/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-kube-api-access-qlng5\") pod \"nova-metadata-0\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.665652 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.926560 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de","Type":"ContainerStarted","Data":"5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf"} Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.926963 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de","Type":"ContainerStarted","Data":"e140144ec0a450c4367b9cfe0053b6718e433805a554f25af15235aaeb9f696d"} Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.927899 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.932628 4946 generic.go:334] "Generic (PLEG): container finished" podID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerID="3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c" exitCode=0 Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.932663 4946 generic.go:334] "Generic (PLEG): container finished" podID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerID="0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85" exitCode=2 Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.932672 4946 generic.go:334] "Generic (PLEG): container finished" podID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerID="e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3" exitCode=0 Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.932725 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerDied","Data":"3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c"} Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.932772 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerDied","Data":"0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85"} Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.932786 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerDied","Data":"e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3"} Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.935050 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="05423f2d-1369-485b-ba07-bcda3b99fa83" containerName="nova-scheduler-scheduler" containerID="cri-o://f8831af5238f203cf08c7dcc6eddea95137d71a86314174f056b6982d075405b" gracePeriod=30 Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.935331 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f69ee496-d325-40e0-81ff-0fb3625a775a","Type":"ContainerStarted","Data":"cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f"} Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.935367 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f69ee496-d325-40e0-81ff-0fb3625a775a","Type":"ContainerStarted","Data":"961d002cc0d977266c82641602e07919bb97e9be177176ffddc7cd5b172c7e24"} Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.935457 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.944076 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.943864038 podStartE2EDuration="2.943864038s" podCreationTimestamp="2025-12-03 07:14:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:33.9409405 +0000 UTC m=+1466.737630609" watchObservedRunningTime="2025-12-03 07:14:33.943864038 +0000 UTC m=+1466.740554147" Dec 03 07:14:33 crc kubenswrapper[4946]: I1203 07:14:33.961609 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.581595351 podStartE2EDuration="2.961594821s" podCreationTimestamp="2025-12-03 07:14:31 +0000 UTC" firstStartedPulling="2025-12-03 07:14:33.046032325 +0000 UTC m=+1465.842722434" lastFinishedPulling="2025-12-03 07:14:33.426031795 +0000 UTC m=+1466.222721904" observedRunningTime="2025-12-03 07:14:33.959588178 +0000 UTC m=+1466.756278287" watchObservedRunningTime="2025-12-03 07:14:33.961594821 +0000 UTC m=+1466.758284930" Dec 03 07:14:34 crc kubenswrapper[4946]: I1203 07:14:34.125456 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:14:34 crc kubenswrapper[4946]: W1203 07:14:34.131850 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3192ad5b_aabb_4fc7_a534_bed599cc6f1c.slice/crio-36302fee7c7aa6579d11f5f9eb6797028eeacbf62be2d66222dbbf9af202d055 WatchSource:0}: Error finding container 36302fee7c7aa6579d11f5f9eb6797028eeacbf62be2d66222dbbf9af202d055: Status 404 returned error can't find the container with id 36302fee7c7aa6579d11f5f9eb6797028eeacbf62be2d66222dbbf9af202d055 Dec 03 07:14:34 crc kubenswrapper[4946]: I1203 07:14:34.945518 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3192ad5b-aabb-4fc7-a534-bed599cc6f1c","Type":"ContainerStarted","Data":"bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3"} Dec 03 07:14:34 crc kubenswrapper[4946]: I1203 07:14:34.945575 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3192ad5b-aabb-4fc7-a534-bed599cc6f1c","Type":"ContainerStarted","Data":"930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6"} Dec 03 07:14:34 crc kubenswrapper[4946]: I1203 07:14:34.945590 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3192ad5b-aabb-4fc7-a534-bed599cc6f1c","Type":"ContainerStarted","Data":"36302fee7c7aa6579d11f5f9eb6797028eeacbf62be2d66222dbbf9af202d055"} Dec 03 07:14:34 crc kubenswrapper[4946]: I1203 07:14:34.975872 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.9758577050000001 podStartE2EDuration="1.975857705s" podCreationTimestamp="2025-12-03 07:14:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:34.972001442 +0000 UTC m=+1467.768691561" watchObservedRunningTime="2025-12-03 07:14:34.975857705 +0000 UTC m=+1467.772547814" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.848684 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.955995 4946 generic.go:334] "Generic (PLEG): container finished" podID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerID="d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5" exitCode=0 Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.956080 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerDied","Data":"d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5"} Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.956100 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.957037 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d","Type":"ContainerDied","Data":"04e9927180229d82e4dfbdc8c049b094f8acfcdfb5729d166858c439350fad86"} Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.957106 4946 scope.go:117] "RemoveContainer" containerID="3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.958375 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-config-data\") pod \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.958412 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-log-httpd\") pod \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.958435 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-scripts\") pod \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.958462 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-sg-core-conf-yaml\") pod \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.958498 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-combined-ca-bundle\") pod \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.958588 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdx5v\" (UniqueName: \"kubernetes.io/projected/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-kube-api-access-mdx5v\") pod \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.958652 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-run-httpd\") pod \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\" (UID: \"ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d\") " Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.958939 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" (UID: "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.959250 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" (UID: "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.959434 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.959457 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.964585 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-kube-api-access-mdx5v" (OuterVolumeSpecName: "kube-api-access-mdx5v") pod "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" (UID: "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d"). InnerVolumeSpecName "kube-api-access-mdx5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.966113 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-scripts" (OuterVolumeSpecName: "scripts") pod "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" (UID: "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:35 crc kubenswrapper[4946]: I1203 07:14:35.991031 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" (UID: "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.034342 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" (UID: "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.061679 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.061708 4946 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.061717 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.061728 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdx5v\" (UniqueName: \"kubernetes.io/projected/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-kube-api-access-mdx5v\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.076854 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-config-data" (OuterVolumeSpecName: "config-data") pod "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" (UID: "ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.156804 4946 scope.go:117] "RemoveContainer" containerID="0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.162872 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.185019 4946 scope.go:117] "RemoveContainer" containerID="d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.202388 4946 scope.go:117] "RemoveContainer" containerID="e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3" Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.211937 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f8831af5238f203cf08c7dcc6eddea95137d71a86314174f056b6982d075405b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.213502 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f8831af5238f203cf08c7dcc6eddea95137d71a86314174f056b6982d075405b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.214586 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f8831af5238f203cf08c7dcc6eddea95137d71a86314174f056b6982d075405b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.214623 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="05423f2d-1369-485b-ba07-bcda3b99fa83" containerName="nova-scheduler-scheduler" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.229005 4946 scope.go:117] "RemoveContainer" containerID="3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c" Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.229366 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c\": container with ID starting with 3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c not found: ID does not exist" containerID="3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.229402 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c"} err="failed to get container status \"3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c\": rpc error: code = NotFound desc = could not find container \"3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c\": container with ID starting with 3b74e712d8505c0e80426ab10cc41e1fe311c0d87f858a0d15816607e3fb612c not found: ID does not exist" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.229427 4946 scope.go:117] "RemoveContainer" containerID="0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85" Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.229708 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85\": container with ID starting with 0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85 not found: ID does not exist" containerID="0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.229766 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85"} err="failed to get container status \"0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85\": rpc error: code = NotFound desc = could not find container \"0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85\": container with ID starting with 0fb251df6c519806c437a826caa208b92ac5ca812179bc09641669054f426c85 not found: ID does not exist" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.229784 4946 scope.go:117] "RemoveContainer" containerID="d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5" Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.230019 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5\": container with ID starting with d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5 not found: ID does not exist" containerID="d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.230048 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5"} err="failed to get container status \"d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5\": rpc error: code = NotFound desc = could not find container \"d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5\": container with ID starting with d6d4783c5b07cadfe1f8d60cdd2ba94c703fc7d7c6f6637aa351e0aa1fae0ad5 not found: ID does not exist" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.230064 4946 scope.go:117] "RemoveContainer" containerID="e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3" Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.230285 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3\": container with ID starting with e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3 not found: ID does not exist" containerID="e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.230313 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3"} err="failed to get container status \"e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3\": rpc error: code = NotFound desc = could not find container \"e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3\": container with ID starting with e038274ebc74c2dcb86448382c2519306c004aa1dacf280bed58b165669ef3a3 not found: ID does not exist" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.301850 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.325061 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.332672 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.333205 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="ceilometer-notification-agent" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.333233 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="ceilometer-notification-agent" Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.333262 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="sg-core" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.333269 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="sg-core" Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.333288 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="proxy-httpd" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.333295 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="proxy-httpd" Dec 03 07:14:36 crc kubenswrapper[4946]: E1203 07:14:36.333322 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="ceilometer-central-agent" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.333329 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="ceilometer-central-agent" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.333495 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="proxy-httpd" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.333508 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="ceilometer-central-agent" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.333518 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="ceilometer-notification-agent" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.333530 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" containerName="sg-core" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.335198 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.339013 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.339276 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.339552 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.340524 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.469962 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.470044 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-log-httpd\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.470086 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-scripts\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.470119 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.470162 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9kc4\" (UniqueName: \"kubernetes.io/projected/c81c7452-0b62-4388-ad47-8901024bb726-kube-api-access-m9kc4\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.470180 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-config-data\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.470216 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.470239 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-run-httpd\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572037 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572103 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-log-httpd\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572133 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-scripts\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572159 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572197 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9kc4\" (UniqueName: \"kubernetes.io/projected/c81c7452-0b62-4388-ad47-8901024bb726-kube-api-access-m9kc4\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572212 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-config-data\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572242 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572259 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-run-httpd\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572669 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-run-httpd\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.572682 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-log-httpd\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.577333 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-config-data\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.578370 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.578545 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-scripts\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.582915 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.584371 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.594195 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9kc4\" (UniqueName: \"kubernetes.io/projected/c81c7452-0b62-4388-ad47-8901024bb726-kube-api-access-m9kc4\") pod \"ceilometer-0\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.652569 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.931887 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.969213 4946 generic.go:334] "Generic (PLEG): container finished" podID="05423f2d-1369-485b-ba07-bcda3b99fa83" containerID="f8831af5238f203cf08c7dcc6eddea95137d71a86314174f056b6982d075405b" exitCode=0 Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.969333 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"05423f2d-1369-485b-ba07-bcda3b99fa83","Type":"ContainerDied","Data":"f8831af5238f203cf08c7dcc6eddea95137d71a86314174f056b6982d075405b"} Dec 03 07:14:36 crc kubenswrapper[4946]: I1203 07:14:36.971058 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerStarted","Data":"2cacce0e7fd2ae758a8bc91995eee38ac3952e6c4b8af40256719b8e710bcf05"} Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.353757 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.493720 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmg72\" (UniqueName: \"kubernetes.io/projected/05423f2d-1369-485b-ba07-bcda3b99fa83-kube-api-access-fmg72\") pod \"05423f2d-1369-485b-ba07-bcda3b99fa83\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.494003 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-config-data\") pod \"05423f2d-1369-485b-ba07-bcda3b99fa83\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.494114 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-combined-ca-bundle\") pod \"05423f2d-1369-485b-ba07-bcda3b99fa83\" (UID: \"05423f2d-1369-485b-ba07-bcda3b99fa83\") " Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.498566 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05423f2d-1369-485b-ba07-bcda3b99fa83-kube-api-access-fmg72" (OuterVolumeSpecName: "kube-api-access-fmg72") pod "05423f2d-1369-485b-ba07-bcda3b99fa83" (UID: "05423f2d-1369-485b-ba07-bcda3b99fa83"). InnerVolumeSpecName "kube-api-access-fmg72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.525231 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05423f2d-1369-485b-ba07-bcda3b99fa83" (UID: "05423f2d-1369-485b-ba07-bcda3b99fa83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.525680 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-config-data" (OuterVolumeSpecName: "config-data") pod "05423f2d-1369-485b-ba07-bcda3b99fa83" (UID: "05423f2d-1369-485b-ba07-bcda3b99fa83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.595827 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmg72\" (UniqueName: \"kubernetes.io/projected/05423f2d-1369-485b-ba07-bcda3b99fa83-kube-api-access-fmg72\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.595859 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.595870 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05423f2d-1369-485b-ba07-bcda3b99fa83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:37 crc kubenswrapper[4946]: I1203 07:14:37.615860 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d" path="/var/lib/kubelet/pods/ecfb79fe-51e7-491b-8e36-4aff7cbe2a8d/volumes" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.013720 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"05423f2d-1369-485b-ba07-bcda3b99fa83","Type":"ContainerDied","Data":"a74941f3509359efe45275672e54a1dab0b82e507460404a7bb257608c495b61"} Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.014032 4946 scope.go:117] "RemoveContainer" containerID="f8831af5238f203cf08c7dcc6eddea95137d71a86314174f056b6982d075405b" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.014169 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.027062 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerStarted","Data":"472b894497c195453df59fead8c87fdb10a93680b5df08ccc87ddf9ae8f84460"} Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.076419 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.091317 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.104404 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:38 crc kubenswrapper[4946]: E1203 07:14:38.105403 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05423f2d-1369-485b-ba07-bcda3b99fa83" containerName="nova-scheduler-scheduler" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.105635 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="05423f2d-1369-485b-ba07-bcda3b99fa83" containerName="nova-scheduler-scheduler" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.106278 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="05423f2d-1369-485b-ba07-bcda3b99fa83" containerName="nova-scheduler-scheduler" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.107640 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.111067 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.113116 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.205595 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-config-data\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.205661 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmn6r\" (UniqueName: \"kubernetes.io/projected/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-kube-api-access-wmn6r\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.205715 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.308537 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-config-data\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.309450 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmn6r\" (UniqueName: \"kubernetes.io/projected/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-kube-api-access-wmn6r\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.309558 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.313454 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.314317 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-config-data\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.331597 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmn6r\" (UniqueName: \"kubernetes.io/projected/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-kube-api-access-wmn6r\") pod \"nova-scheduler-0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.464726 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.666319 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.667266 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.919764 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:14:38 crc kubenswrapper[4946]: W1203 07:14:38.931011 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod341d1e5a_8309_4cb1_8f50_2ddf556a7db0.slice/crio-d24f3d40c3988ac37ee76502def45d5ad761af54bc70084746041b2b89f8a38a WatchSource:0}: Error finding container d24f3d40c3988ac37ee76502def45d5ad761af54bc70084746041b2b89f8a38a: Status 404 returned error can't find the container with id d24f3d40c3988ac37ee76502def45d5ad761af54bc70084746041b2b89f8a38a Dec 03 07:14:38 crc kubenswrapper[4946]: I1203 07:14:38.932256 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.021717 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5ptk\" (UniqueName: \"kubernetes.io/projected/7d1d5102-9075-4423-8735-41834caa451a-kube-api-access-q5ptk\") pod \"7d1d5102-9075-4423-8735-41834caa451a\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.022119 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-config-data\") pod \"7d1d5102-9075-4423-8735-41834caa451a\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.022142 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-combined-ca-bundle\") pod \"7d1d5102-9075-4423-8735-41834caa451a\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.022236 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d1d5102-9075-4423-8735-41834caa451a-logs\") pod \"7d1d5102-9075-4423-8735-41834caa451a\" (UID: \"7d1d5102-9075-4423-8735-41834caa451a\") " Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.037488 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d1d5102-9075-4423-8735-41834caa451a-logs" (OuterVolumeSpecName: "logs") pod "7d1d5102-9075-4423-8735-41834caa451a" (UID: "7d1d5102-9075-4423-8735-41834caa451a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.040720 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d1d5102-9075-4423-8735-41834caa451a-kube-api-access-q5ptk" (OuterVolumeSpecName: "kube-api-access-q5ptk") pod "7d1d5102-9075-4423-8735-41834caa451a" (UID: "7d1d5102-9075-4423-8735-41834caa451a"). InnerVolumeSpecName "kube-api-access-q5ptk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.078600 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-config-data" (OuterVolumeSpecName: "config-data") pod "7d1d5102-9075-4423-8735-41834caa451a" (UID: "7d1d5102-9075-4423-8735-41834caa451a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.081945 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d1d5102-9075-4423-8735-41834caa451a" (UID: "7d1d5102-9075-4423-8735-41834caa451a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.091525 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerStarted","Data":"d01f7981c5efe052b4cf6101ea504e6ce8ea707e982a9081e7754f677063658e"} Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.094305 4946 generic.go:334] "Generic (PLEG): container finished" podID="7d1d5102-9075-4423-8735-41834caa451a" containerID="91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551" exitCode=0 Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.094442 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d1d5102-9075-4423-8735-41834caa451a","Type":"ContainerDied","Data":"91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551"} Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.094513 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d1d5102-9075-4423-8735-41834caa451a","Type":"ContainerDied","Data":"e4ccfa04a8d18a7c9b28c5c6632f6f4b8d7f9c64fb7f01cd1768cb4666a5cd0c"} Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.094586 4946 scope.go:117] "RemoveContainer" containerID="91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.094786 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.096578 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"341d1e5a-8309-4cb1-8f50-2ddf556a7db0","Type":"ContainerStarted","Data":"d24f3d40c3988ac37ee76502def45d5ad761af54bc70084746041b2b89f8a38a"} Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.131346 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d1d5102-9075-4423-8735-41834caa451a-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.131402 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5ptk\" (UniqueName: \"kubernetes.io/projected/7d1d5102-9075-4423-8735-41834caa451a-kube-api-access-q5ptk\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.131415 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.131428 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d1d5102-9075-4423-8735-41834caa451a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.172183 4946 scope.go:117] "RemoveContainer" containerID="7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.199245 4946 scope.go:117] "RemoveContainer" containerID="91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551" Dec 03 07:14:39 crc kubenswrapper[4946]: E1203 07:14:39.200119 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551\": container with ID starting with 91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551 not found: ID does not exist" containerID="91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.200161 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551"} err="failed to get container status \"91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551\": rpc error: code = NotFound desc = could not find container \"91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551\": container with ID starting with 91d507574ffe5ef4437a5b0c410534f6e238499b69b862a2dc38c3c96f166551 not found: ID does not exist" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.200189 4946 scope.go:117] "RemoveContainer" containerID="7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f" Dec 03 07:14:39 crc kubenswrapper[4946]: E1203 07:14:39.200571 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f\": container with ID starting with 7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f not found: ID does not exist" containerID="7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.200593 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f"} err="failed to get container status \"7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f\": rpc error: code = NotFound desc = could not find container \"7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f\": container with ID starting with 7b41504daea599b44e23559b30d5402e1895ff141f62dc4a0d37acfee3fd336f not found: ID does not exist" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.209933 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.233339 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.242834 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:39 crc kubenswrapper[4946]: E1203 07:14:39.243325 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-api" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.243391 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-api" Dec 03 07:14:39 crc kubenswrapper[4946]: E1203 07:14:39.243473 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-log" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.243531 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-log" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.243777 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-log" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.243877 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d1d5102-9075-4423-8735-41834caa451a" containerName="nova-api-api" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.244899 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.251013 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.273187 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.334594 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-config-data\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.334729 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.334800 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c68706-049c-4550-ac86-3982b9d0eb70-logs\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.334855 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmczj\" (UniqueName: \"kubernetes.io/projected/44c68706-049c-4550-ac86-3982b9d0eb70-kube-api-access-qmczj\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.436568 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.436944 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c68706-049c-4550-ac86-3982b9d0eb70-logs\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.437004 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmczj\" (UniqueName: \"kubernetes.io/projected/44c68706-049c-4550-ac86-3982b9d0eb70-kube-api-access-qmczj\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.437024 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-config-data\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.437485 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c68706-049c-4550-ac86-3982b9d0eb70-logs\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.445622 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-config-data\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.460341 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.464586 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmczj\" (UniqueName: \"kubernetes.io/projected/44c68706-049c-4550-ac86-3982b9d0eb70-kube-api-access-qmczj\") pod \"nova-api-0\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.591031 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.615410 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05423f2d-1369-485b-ba07-bcda3b99fa83" path="/var/lib/kubelet/pods/05423f2d-1369-485b-ba07-bcda3b99fa83/volumes" Dec 03 07:14:39 crc kubenswrapper[4946]: I1203 07:14:39.616597 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d1d5102-9075-4423-8735-41834caa451a" path="/var/lib/kubelet/pods/7d1d5102-9075-4423-8735-41834caa451a/volumes" Dec 03 07:14:40 crc kubenswrapper[4946]: I1203 07:14:40.083870 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:14:40 crc kubenswrapper[4946]: W1203 07:14:40.108758 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44c68706_049c_4550_ac86_3982b9d0eb70.slice/crio-7b7fb931bb467b00981514154c80e1c67634cc8340575413b4934a0ea41e14ec WatchSource:0}: Error finding container 7b7fb931bb467b00981514154c80e1c67634cc8340575413b4934a0ea41e14ec: Status 404 returned error can't find the container with id 7b7fb931bb467b00981514154c80e1c67634cc8340575413b4934a0ea41e14ec Dec 03 07:14:40 crc kubenswrapper[4946]: I1203 07:14:40.132088 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerStarted","Data":"c6a3e5b8e2f5447abcb8ac293e3ff5788a44666315066833a7c030a378495ed9"} Dec 03 07:14:40 crc kubenswrapper[4946]: I1203 07:14:40.135841 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"341d1e5a-8309-4cb1-8f50-2ddf556a7db0","Type":"ContainerStarted","Data":"4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314"} Dec 03 07:14:40 crc kubenswrapper[4946]: I1203 07:14:40.156660 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.156641764 podStartE2EDuration="2.156641764s" podCreationTimestamp="2025-12-03 07:14:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:40.153391067 +0000 UTC m=+1472.950081186" watchObservedRunningTime="2025-12-03 07:14:40.156641764 +0000 UTC m=+1472.953331883" Dec 03 07:14:41 crc kubenswrapper[4946]: I1203 07:14:41.149197 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c68706-049c-4550-ac86-3982b9d0eb70","Type":"ContainerStarted","Data":"7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949"} Dec 03 07:14:41 crc kubenswrapper[4946]: I1203 07:14:41.149708 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c68706-049c-4550-ac86-3982b9d0eb70","Type":"ContainerStarted","Data":"9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038"} Dec 03 07:14:41 crc kubenswrapper[4946]: I1203 07:14:41.149718 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c68706-049c-4550-ac86-3982b9d0eb70","Type":"ContainerStarted","Data":"7b7fb931bb467b00981514154c80e1c67634cc8340575413b4934a0ea41e14ec"} Dec 03 07:14:41 crc kubenswrapper[4946]: I1203 07:14:41.154056 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerStarted","Data":"c38091b17a25c86f76292d47faafe82eaf48ed142e05937e9350d421568f4d36"} Dec 03 07:14:41 crc kubenswrapper[4946]: I1203 07:14:41.154457 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 07:14:41 crc kubenswrapper[4946]: I1203 07:14:41.197965 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.6118487369999999 podStartE2EDuration="5.197946009s" podCreationTimestamp="2025-12-03 07:14:36 +0000 UTC" firstStartedPulling="2025-12-03 07:14:36.937585316 +0000 UTC m=+1469.734275435" lastFinishedPulling="2025-12-03 07:14:40.523682608 +0000 UTC m=+1473.320372707" observedRunningTime="2025-12-03 07:14:41.194915078 +0000 UTC m=+1473.991605197" watchObservedRunningTime="2025-12-03 07:14:41.197946009 +0000 UTC m=+1473.994636118" Dec 03 07:14:41 crc kubenswrapper[4946]: I1203 07:14:41.199816 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.199808429 podStartE2EDuration="2.199808429s" podCreationTimestamp="2025-12-03 07:14:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:41.172424497 +0000 UTC m=+1473.969114646" watchObservedRunningTime="2025-12-03 07:14:41.199808429 +0000 UTC m=+1473.996498538" Dec 03 07:14:42 crc kubenswrapper[4946]: I1203 07:14:42.272611 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 07:14:42 crc kubenswrapper[4946]: I1203 07:14:42.354402 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 07:14:43 crc kubenswrapper[4946]: I1203 07:14:43.465833 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 07:14:43 crc kubenswrapper[4946]: I1203 07:14:43.666704 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 07:14:43 crc kubenswrapper[4946]: I1203 07:14:43.666759 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 07:14:44 crc kubenswrapper[4946]: I1203 07:14:44.681961 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 07:14:44 crc kubenswrapper[4946]: I1203 07:14:44.682024 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 07:14:48 crc kubenswrapper[4946]: I1203 07:14:48.465375 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 07:14:48 crc kubenswrapper[4946]: I1203 07:14:48.508442 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 07:14:49 crc kubenswrapper[4946]: I1203 07:14:49.293103 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 07:14:49 crc kubenswrapper[4946]: I1203 07:14:49.591332 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 07:14:49 crc kubenswrapper[4946]: I1203 07:14:49.591405 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 07:14:50 crc kubenswrapper[4946]: I1203 07:14:50.673955 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 07:14:50 crc kubenswrapper[4946]: I1203 07:14:50.673957 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 07:14:53 crc kubenswrapper[4946]: I1203 07:14:53.677017 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 07:14:53 crc kubenswrapper[4946]: I1203 07:14:53.679093 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 07:14:53 crc kubenswrapper[4946]: I1203 07:14:53.685728 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 07:14:54 crc kubenswrapper[4946]: I1203 07:14:54.313875 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.210084 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.215305 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-config-data\") pod \"e4962658-f138-4157-8ec0-a4d5d951711d\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.215386 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-combined-ca-bundle\") pod \"e4962658-f138-4157-8ec0-a4d5d951711d\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.215554 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfvfk\" (UniqueName: \"kubernetes.io/projected/e4962658-f138-4157-8ec0-a4d5d951711d-kube-api-access-kfvfk\") pod \"e4962658-f138-4157-8ec0-a4d5d951711d\" (UID: \"e4962658-f138-4157-8ec0-a4d5d951711d\") " Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.224019 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4962658-f138-4157-8ec0-a4d5d951711d-kube-api-access-kfvfk" (OuterVolumeSpecName: "kube-api-access-kfvfk") pod "e4962658-f138-4157-8ec0-a4d5d951711d" (UID: "e4962658-f138-4157-8ec0-a4d5d951711d"). InnerVolumeSpecName "kube-api-access-kfvfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.264300 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-config-data" (OuterVolumeSpecName: "config-data") pod "e4962658-f138-4157-8ec0-a4d5d951711d" (UID: "e4962658-f138-4157-8ec0-a4d5d951711d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.279956 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4962658-f138-4157-8ec0-a4d5d951711d" (UID: "e4962658-f138-4157-8ec0-a4d5d951711d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.316950 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.316981 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4962658-f138-4157-8ec0-a4d5d951711d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.316992 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfvfk\" (UniqueName: \"kubernetes.io/projected/e4962658-f138-4157-8ec0-a4d5d951711d-kube-api-access-kfvfk\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.329641 4946 generic.go:334] "Generic (PLEG): container finished" podID="e4962658-f138-4157-8ec0-a4d5d951711d" containerID="6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df" exitCode=137 Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.329712 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.329777 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e4962658-f138-4157-8ec0-a4d5d951711d","Type":"ContainerDied","Data":"6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df"} Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.329816 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e4962658-f138-4157-8ec0-a4d5d951711d","Type":"ContainerDied","Data":"d56766efee9d3aa1ec20fb1da046a1b9f36c4211f24bb0cdd980c79cea062768"} Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.329842 4946 scope.go:117] "RemoveContainer" containerID="6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.373703 4946 scope.go:117] "RemoveContainer" containerID="6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df" Dec 03 07:14:56 crc kubenswrapper[4946]: E1203 07:14:56.374126 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df\": container with ID starting with 6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df not found: ID does not exist" containerID="6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.374150 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df"} err="failed to get container status \"6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df\": rpc error: code = NotFound desc = could not find container \"6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df\": container with ID starting with 6277a6e7c12d15aa37f0a7e08db69de6639fd795cd5eb42e954935fb8f7db2df not found: ID does not exist" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.383893 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.401483 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.415543 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:56 crc kubenswrapper[4946]: E1203 07:14:56.416197 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4962658-f138-4157-8ec0-a4d5d951711d" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.416225 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4962658-f138-4157-8ec0-a4d5d951711d" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.416564 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4962658-f138-4157-8ec0-a4d5d951711d" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.417615 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.420493 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.421076 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.427635 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.431970 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.523791 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d79p9\" (UniqueName: \"kubernetes.io/projected/04384280-d303-475d-8abc-c9c957db7fd9-kube-api-access-d79p9\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.524154 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.524220 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.524344 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.524669 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.626753 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.626867 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.626899 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d79p9\" (UniqueName: \"kubernetes.io/projected/04384280-d303-475d-8abc-c9c957db7fd9-kube-api-access-d79p9\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.626975 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.626994 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.631797 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.632280 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.633259 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.637434 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.641370 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d79p9\" (UniqueName: \"kubernetes.io/projected/04384280-d303-475d-8abc-c9c957db7fd9-kube-api-access-d79p9\") pod \"nova-cell1-novncproxy-0\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:56 crc kubenswrapper[4946]: I1203 07:14:56.737850 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:14:57 crc kubenswrapper[4946]: I1203 07:14:57.309616 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:14:57 crc kubenswrapper[4946]: I1203 07:14:57.347956 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"04384280-d303-475d-8abc-c9c957db7fd9","Type":"ContainerStarted","Data":"e17dd66e133e84e0005889ea0db01e05d4c1c7e86dec7c296cea1a9ac8c3a058"} Dec 03 07:14:57 crc kubenswrapper[4946]: I1203 07:14:57.611005 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4962658-f138-4157-8ec0-a4d5d951711d" path="/var/lib/kubelet/pods/e4962658-f138-4157-8ec0-a4d5d951711d/volumes" Dec 03 07:14:58 crc kubenswrapper[4946]: I1203 07:14:58.399899 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"04384280-d303-475d-8abc-c9c957db7fd9","Type":"ContainerStarted","Data":"c1b84099e3ff7baff4b0d71a61709731ec4d0a9f91be8cac1668b867cdc1f2bf"} Dec 03 07:14:58 crc kubenswrapper[4946]: I1203 07:14:58.423987 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.423966432 podStartE2EDuration="2.423966432s" podCreationTimestamp="2025-12-03 07:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:14:58.421581908 +0000 UTC m=+1491.218272057" watchObservedRunningTime="2025-12-03 07:14:58.423966432 +0000 UTC m=+1491.220656551" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.612579 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.613015 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.614200 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.614259 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.618024 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.619725 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.825680 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c8fb5597c-2d82q"] Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.827507 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.865051 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8fb5597c-2d82q"] Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.913932 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-svc\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.913979 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.914033 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-config\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.914145 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2skq\" (UniqueName: \"kubernetes.io/projected/8126945c-b3f3-4fc8-a5eb-553c41a439d5-kube-api-access-j2skq\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.914225 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:14:59 crc kubenswrapper[4946]: I1203 07:14:59.914297 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.015667 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-config\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.015788 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2skq\" (UniqueName: \"kubernetes.io/projected/8126945c-b3f3-4fc8-a5eb-553c41a439d5-kube-api-access-j2skq\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.015888 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.015919 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.015950 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-svc\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.015976 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.016813 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.017318 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-config\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.017800 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.017948 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-svc\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.018123 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.039431 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2skq\" (UniqueName: \"kubernetes.io/projected/8126945c-b3f3-4fc8-a5eb-553c41a439d5-kube-api-access-j2skq\") pod \"dnsmasq-dns-5c8fb5597c-2d82q\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.142819 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw"] Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.144269 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.147011 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.148339 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.149102 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.158874 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw"] Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.218563 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0e9ec0-504f-496e-b9e5-c7aa89352318-config-volume\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.218650 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25swh\" (UniqueName: \"kubernetes.io/projected/af0e9ec0-504f-496e-b9e5-c7aa89352318-kube-api-access-25swh\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.218693 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0e9ec0-504f-496e-b9e5-c7aa89352318-secret-volume\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.320611 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0e9ec0-504f-496e-b9e5-c7aa89352318-config-volume\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.321314 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25swh\" (UniqueName: \"kubernetes.io/projected/af0e9ec0-504f-496e-b9e5-c7aa89352318-kube-api-access-25swh\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.321350 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0e9ec0-504f-496e-b9e5-c7aa89352318-secret-volume\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.322916 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0e9ec0-504f-496e-b9e5-c7aa89352318-config-volume\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.328105 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0e9ec0-504f-496e-b9e5-c7aa89352318-secret-volume\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.343586 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25swh\" (UniqueName: \"kubernetes.io/projected/af0e9ec0-504f-496e-b9e5-c7aa89352318-kube-api-access-25swh\") pod \"collect-profiles-29412435-b9xpw\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.467762 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.639862 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8fb5597c-2d82q"] Dec 03 07:15:00 crc kubenswrapper[4946]: W1203 07:15:00.937569 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf0e9ec0_504f_496e_b9e5_c7aa89352318.slice/crio-5fbb9b9a00dcf41401183bd29a32fbcde5ded44bdf10a78b99ed4445d2042241 WatchSource:0}: Error finding container 5fbb9b9a00dcf41401183bd29a32fbcde5ded44bdf10a78b99ed4445d2042241: Status 404 returned error can't find the container with id 5fbb9b9a00dcf41401183bd29a32fbcde5ded44bdf10a78b99ed4445d2042241 Dec 03 07:15:00 crc kubenswrapper[4946]: I1203 07:15:00.940518 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw"] Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.452103 4946 generic.go:334] "Generic (PLEG): container finished" podID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerID="755c77d55195689162b2c7aae3960309c3338cb7b0ded20090fed4b7656c4494" exitCode=0 Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.452278 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" event={"ID":"8126945c-b3f3-4fc8-a5eb-553c41a439d5","Type":"ContainerDied","Data":"755c77d55195689162b2c7aae3960309c3338cb7b0ded20090fed4b7656c4494"} Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.452568 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" event={"ID":"8126945c-b3f3-4fc8-a5eb-553c41a439d5","Type":"ContainerStarted","Data":"b9ad8d184d12d742da68ba2d1c8fd01b68c215fe4f2e6d11c071a45476078824"} Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.454687 4946 generic.go:334] "Generic (PLEG): container finished" podID="af0e9ec0-504f-496e-b9e5-c7aa89352318" containerID="59b2e84655c7126bb52221921ca30a70758edc78fac3c60480dd87b24524d12a" exitCode=0 Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.455338 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" event={"ID":"af0e9ec0-504f-496e-b9e5-c7aa89352318","Type":"ContainerDied","Data":"59b2e84655c7126bb52221921ca30a70758edc78fac3c60480dd87b24524d12a"} Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.455371 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" event={"ID":"af0e9ec0-504f-496e-b9e5-c7aa89352318","Type":"ContainerStarted","Data":"5fbb9b9a00dcf41401183bd29a32fbcde5ded44bdf10a78b99ed4445d2042241"} Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.738545 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.827360 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.827639 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="ceilometer-central-agent" containerID="cri-o://472b894497c195453df59fead8c87fdb10a93680b5df08ccc87ddf9ae8f84460" gracePeriod=30 Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.827711 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="sg-core" containerID="cri-o://c6a3e5b8e2f5447abcb8ac293e3ff5788a44666315066833a7c030a378495ed9" gracePeriod=30 Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.827761 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="ceilometer-notification-agent" containerID="cri-o://d01f7981c5efe052b4cf6101ea504e6ce8ea707e982a9081e7754f677063658e" gracePeriod=30 Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.827790 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="proxy-httpd" containerID="cri-o://c38091b17a25c86f76292d47faafe82eaf48ed142e05937e9350d421568f4d36" gracePeriod=30 Dec 03 07:15:01 crc kubenswrapper[4946]: I1203 07:15:01.871512 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.192:3000/\": read tcp 10.217.0.2:49446->10.217.0.192:3000: read: connection reset by peer" Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.465420 4946 generic.go:334] "Generic (PLEG): container finished" podID="c81c7452-0b62-4388-ad47-8901024bb726" containerID="c38091b17a25c86f76292d47faafe82eaf48ed142e05937e9350d421568f4d36" exitCode=0 Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.465450 4946 generic.go:334] "Generic (PLEG): container finished" podID="c81c7452-0b62-4388-ad47-8901024bb726" containerID="c6a3e5b8e2f5447abcb8ac293e3ff5788a44666315066833a7c030a378495ed9" exitCode=2 Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.465456 4946 generic.go:334] "Generic (PLEG): container finished" podID="c81c7452-0b62-4388-ad47-8901024bb726" containerID="472b894497c195453df59fead8c87fdb10a93680b5df08ccc87ddf9ae8f84460" exitCode=0 Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.465515 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerDied","Data":"c38091b17a25c86f76292d47faafe82eaf48ed142e05937e9350d421568f4d36"} Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.465585 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerDied","Data":"c6a3e5b8e2f5447abcb8ac293e3ff5788a44666315066833a7c030a378495ed9"} Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.465611 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerDied","Data":"472b894497c195453df59fead8c87fdb10a93680b5df08ccc87ddf9ae8f84460"} Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.468097 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" event={"ID":"8126945c-b3f3-4fc8-a5eb-553c41a439d5","Type":"ContainerStarted","Data":"b376d6f31b26fb3b5d8428d35df403e2d7ba33125152c97ad1261520a5e2f273"} Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.468322 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.491686 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" podStartSLOduration=3.491670704 podStartE2EDuration="3.491670704s" podCreationTimestamp="2025-12-03 07:14:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:15:02.48668136 +0000 UTC m=+1495.283371479" watchObservedRunningTime="2025-12-03 07:15:02.491670704 +0000 UTC m=+1495.288360813" Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.803827 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.804381 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-log" containerID="cri-o://9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038" gracePeriod=30 Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.804515 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-api" containerID="cri-o://7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949" gracePeriod=30 Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.858898 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.977786 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0e9ec0-504f-496e-b9e5-c7aa89352318-config-volume\") pod \"af0e9ec0-504f-496e-b9e5-c7aa89352318\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.977889 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25swh\" (UniqueName: \"kubernetes.io/projected/af0e9ec0-504f-496e-b9e5-c7aa89352318-kube-api-access-25swh\") pod \"af0e9ec0-504f-496e-b9e5-c7aa89352318\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.978042 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0e9ec0-504f-496e-b9e5-c7aa89352318-secret-volume\") pod \"af0e9ec0-504f-496e-b9e5-c7aa89352318\" (UID: \"af0e9ec0-504f-496e-b9e5-c7aa89352318\") " Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.978544 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0e9ec0-504f-496e-b9e5-c7aa89352318-config-volume" (OuterVolumeSpecName: "config-volume") pod "af0e9ec0-504f-496e-b9e5-c7aa89352318" (UID: "af0e9ec0-504f-496e-b9e5-c7aa89352318"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.983712 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0e9ec0-504f-496e-b9e5-c7aa89352318-kube-api-access-25swh" (OuterVolumeSpecName: "kube-api-access-25swh") pod "af0e9ec0-504f-496e-b9e5-c7aa89352318" (UID: "af0e9ec0-504f-496e-b9e5-c7aa89352318"). InnerVolumeSpecName "kube-api-access-25swh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:02 crc kubenswrapper[4946]: I1203 07:15:02.983805 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af0e9ec0-504f-496e-b9e5-c7aa89352318-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "af0e9ec0-504f-496e-b9e5-c7aa89352318" (UID: "af0e9ec0-504f-496e-b9e5-c7aa89352318"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:03 crc kubenswrapper[4946]: I1203 07:15:03.079575 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0e9ec0-504f-496e-b9e5-c7aa89352318-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:03 crc kubenswrapper[4946]: I1203 07:15:03.079603 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0e9ec0-504f-496e-b9e5-c7aa89352318-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:03 crc kubenswrapper[4946]: I1203 07:15:03.079614 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25swh\" (UniqueName: \"kubernetes.io/projected/af0e9ec0-504f-496e-b9e5-c7aa89352318-kube-api-access-25swh\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:03 crc kubenswrapper[4946]: I1203 07:15:03.477153 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" event={"ID":"af0e9ec0-504f-496e-b9e5-c7aa89352318","Type":"ContainerDied","Data":"5fbb9b9a00dcf41401183bd29a32fbcde5ded44bdf10a78b99ed4445d2042241"} Dec 03 07:15:03 crc kubenswrapper[4946]: I1203 07:15:03.477184 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fbb9b9a00dcf41401183bd29a32fbcde5ded44bdf10a78b99ed4445d2042241" Dec 03 07:15:03 crc kubenswrapper[4946]: I1203 07:15:03.477242 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw" Dec 03 07:15:03 crc kubenswrapper[4946]: I1203 07:15:03.484852 4946 generic.go:334] "Generic (PLEG): container finished" podID="44c68706-049c-4550-ac86-3982b9d0eb70" containerID="9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038" exitCode=143 Dec 03 07:15:03 crc kubenswrapper[4946]: I1203 07:15:03.485470 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c68706-049c-4550-ac86-3982b9d0eb70","Type":"ContainerDied","Data":"9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038"} Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.495003 4946 generic.go:334] "Generic (PLEG): container finished" podID="c81c7452-0b62-4388-ad47-8901024bb726" containerID="d01f7981c5efe052b4cf6101ea504e6ce8ea707e982a9081e7754f677063658e" exitCode=0 Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.495076 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerDied","Data":"d01f7981c5efe052b4cf6101ea504e6ce8ea707e982a9081e7754f677063658e"} Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.495358 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c81c7452-0b62-4388-ad47-8901024bb726","Type":"ContainerDied","Data":"2cacce0e7fd2ae758a8bc91995eee38ac3952e6c4b8af40256719b8e710bcf05"} Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.495370 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cacce0e7fd2ae758a8bc91995eee38ac3952e6c4b8af40256719b8e710bcf05" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.560058 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.726481 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-combined-ca-bundle\") pod \"c81c7452-0b62-4388-ad47-8901024bb726\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.726527 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-run-httpd\") pod \"c81c7452-0b62-4388-ad47-8901024bb726\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.726561 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-config-data\") pod \"c81c7452-0b62-4388-ad47-8901024bb726\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.726637 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-scripts\") pod \"c81c7452-0b62-4388-ad47-8901024bb726\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.726676 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-sg-core-conf-yaml\") pod \"c81c7452-0b62-4388-ad47-8901024bb726\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.726707 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9kc4\" (UniqueName: \"kubernetes.io/projected/c81c7452-0b62-4388-ad47-8901024bb726-kube-api-access-m9kc4\") pod \"c81c7452-0b62-4388-ad47-8901024bb726\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.726758 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-log-httpd\") pod \"c81c7452-0b62-4388-ad47-8901024bb726\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.726848 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-ceilometer-tls-certs\") pod \"c81c7452-0b62-4388-ad47-8901024bb726\" (UID: \"c81c7452-0b62-4388-ad47-8901024bb726\") " Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.727315 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c81c7452-0b62-4388-ad47-8901024bb726" (UID: "c81c7452-0b62-4388-ad47-8901024bb726"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.727531 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c81c7452-0b62-4388-ad47-8901024bb726" (UID: "c81c7452-0b62-4388-ad47-8901024bb726"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.732515 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-scripts" (OuterVolumeSpecName: "scripts") pod "c81c7452-0b62-4388-ad47-8901024bb726" (UID: "c81c7452-0b62-4388-ad47-8901024bb726"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.734950 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c81c7452-0b62-4388-ad47-8901024bb726-kube-api-access-m9kc4" (OuterVolumeSpecName: "kube-api-access-m9kc4") pod "c81c7452-0b62-4388-ad47-8901024bb726" (UID: "c81c7452-0b62-4388-ad47-8901024bb726"). InnerVolumeSpecName "kube-api-access-m9kc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.753345 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c81c7452-0b62-4388-ad47-8901024bb726" (UID: "c81c7452-0b62-4388-ad47-8901024bb726"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.800159 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c81c7452-0b62-4388-ad47-8901024bb726" (UID: "c81c7452-0b62-4388-ad47-8901024bb726"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.824264 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c81c7452-0b62-4388-ad47-8901024bb726" (UID: "c81c7452-0b62-4388-ad47-8901024bb726"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.829351 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.829383 4946 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.829395 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9kc4\" (UniqueName: \"kubernetes.io/projected/c81c7452-0b62-4388-ad47-8901024bb726-kube-api-access-m9kc4\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.829405 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.829414 4946 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.829424 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.829432 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c81c7452-0b62-4388-ad47-8901024bb726-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.869965 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-config-data" (OuterVolumeSpecName: "config-data") pod "c81c7452-0b62-4388-ad47-8901024bb726" (UID: "c81c7452-0b62-4388-ad47-8901024bb726"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:04 crc kubenswrapper[4946]: I1203 07:15:04.930893 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c81c7452-0b62-4388-ad47-8901024bb726-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.505219 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.549926 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.560797 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.608891 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c81c7452-0b62-4388-ad47-8901024bb726" path="/var/lib/kubelet/pods/c81c7452-0b62-4388-ad47-8901024bb726/volumes" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.610014 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:15:05 crc kubenswrapper[4946]: E1203 07:15:05.610433 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="proxy-httpd" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.610465 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="proxy-httpd" Dec 03 07:15:05 crc kubenswrapper[4946]: E1203 07:15:05.610499 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="sg-core" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.610511 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="sg-core" Dec 03 07:15:05 crc kubenswrapper[4946]: E1203 07:15:05.610529 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="ceilometer-notification-agent" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.610539 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="ceilometer-notification-agent" Dec 03 07:15:05 crc kubenswrapper[4946]: E1203 07:15:05.610569 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="ceilometer-central-agent" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.610581 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="ceilometer-central-agent" Dec 03 07:15:05 crc kubenswrapper[4946]: E1203 07:15:05.610609 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0e9ec0-504f-496e-b9e5-c7aa89352318" containerName="collect-profiles" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.610620 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0e9ec0-504f-496e-b9e5-c7aa89352318" containerName="collect-profiles" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.611332 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="ceilometer-central-agent" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.611367 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="proxy-httpd" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.611382 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0e9ec0-504f-496e-b9e5-c7aa89352318" containerName="collect-profiles" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.611397 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="sg-core" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.611431 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c81c7452-0b62-4388-ad47-8901024bb726" containerName="ceilometer-notification-agent" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.613840 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.621838 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.622514 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.647502 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.647584 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-config-data\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.647655 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-scripts\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.647700 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.647781 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvn6t\" (UniqueName: \"kubernetes.io/projected/d9d0798c-6a8a-4df3-9270-c75d49a2379c-kube-api-access-vvn6t\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.647839 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-log-httpd\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.647884 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-run-httpd\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.647923 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.669955 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.672791 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.749267 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.749341 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-config-data\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.749391 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-scripts\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.749418 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.749460 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvn6t\" (UniqueName: \"kubernetes.io/projected/d9d0798c-6a8a-4df3-9270-c75d49a2379c-kube-api-access-vvn6t\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.749499 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-log-httpd\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.749535 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-run-httpd\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.749564 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.751362 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-log-httpd\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.751454 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-run-httpd\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.755967 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.756185 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-scripts\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.759578 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-config-data\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.766394 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.768685 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:05 crc kubenswrapper[4946]: I1203 07:15:05.772390 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvn6t\" (UniqueName: \"kubernetes.io/projected/d9d0798c-6a8a-4df3-9270-c75d49a2379c-kube-api-access-vvn6t\") pod \"ceilometer-0\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " pod="openstack/ceilometer-0" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.000264 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.463690 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.516163 4946 generic.go:334] "Generic (PLEG): container finished" podID="44c68706-049c-4550-ac86-3982b9d0eb70" containerID="7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949" exitCode=0 Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.516202 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c68706-049c-4550-ac86-3982b9d0eb70","Type":"ContainerDied","Data":"7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949"} Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.516226 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c68706-049c-4550-ac86-3982b9d0eb70","Type":"ContainerDied","Data":"7b7fb931bb467b00981514154c80e1c67634cc8340575413b4934a0ea41e14ec"} Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.516243 4946 scope.go:117] "RemoveContainer" containerID="7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.516348 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.572652 4946 scope.go:117] "RemoveContainer" containerID="9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.587938 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.588844 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-config-data\") pod \"44c68706-049c-4550-ac86-3982b9d0eb70\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.588990 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmczj\" (UniqueName: \"kubernetes.io/projected/44c68706-049c-4550-ac86-3982b9d0eb70-kube-api-access-qmczj\") pod \"44c68706-049c-4550-ac86-3982b9d0eb70\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.589079 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c68706-049c-4550-ac86-3982b9d0eb70-logs\") pod \"44c68706-049c-4550-ac86-3982b9d0eb70\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.589096 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-combined-ca-bundle\") pod \"44c68706-049c-4550-ac86-3982b9d0eb70\" (UID: \"44c68706-049c-4550-ac86-3982b9d0eb70\") " Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.590746 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c68706-049c-4550-ac86-3982b9d0eb70-logs" (OuterVolumeSpecName: "logs") pod "44c68706-049c-4550-ac86-3982b9d0eb70" (UID: "44c68706-049c-4550-ac86-3982b9d0eb70"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.596060 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c68706-049c-4550-ac86-3982b9d0eb70-kube-api-access-qmczj" (OuterVolumeSpecName: "kube-api-access-qmczj") pod "44c68706-049c-4550-ac86-3982b9d0eb70" (UID: "44c68706-049c-4550-ac86-3982b9d0eb70"). InnerVolumeSpecName "kube-api-access-qmczj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:06 crc kubenswrapper[4946]: W1203 07:15:06.615986 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9d0798c_6a8a_4df3_9270_c75d49a2379c.slice/crio-c98628a377aedb34fb875b92a9ac07a99529b16c880d351880505ad04c04f15c WatchSource:0}: Error finding container c98628a377aedb34fb875b92a9ac07a99529b16c880d351880505ad04c04f15c: Status 404 returned error can't find the container with id c98628a377aedb34fb875b92a9ac07a99529b16c880d351880505ad04c04f15c Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.625454 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44c68706-049c-4550-ac86-3982b9d0eb70" (UID: "44c68706-049c-4550-ac86-3982b9d0eb70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.670630 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-config-data" (OuterVolumeSpecName: "config-data") pod "44c68706-049c-4550-ac86-3982b9d0eb70" (UID: "44c68706-049c-4550-ac86-3982b9d0eb70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.691545 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c68706-049c-4550-ac86-3982b9d0eb70-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.691575 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.691585 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c68706-049c-4550-ac86-3982b9d0eb70-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.691596 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmczj\" (UniqueName: \"kubernetes.io/projected/44c68706-049c-4550-ac86-3982b9d0eb70-kube-api-access-qmczj\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.739221 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.751721 4946 scope.go:117] "RemoveContainer" containerID="7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949" Dec 03 07:15:06 crc kubenswrapper[4946]: E1203 07:15:06.752114 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949\": container with ID starting with 7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949 not found: ID does not exist" containerID="7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.752145 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949"} err="failed to get container status \"7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949\": rpc error: code = NotFound desc = could not find container \"7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949\": container with ID starting with 7e9c47ecdc766470cf7ed75b3db4f1ca3f81a6c8efc9545729916eb35b54b949 not found: ID does not exist" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.752164 4946 scope.go:117] "RemoveContainer" containerID="9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038" Dec 03 07:15:06 crc kubenswrapper[4946]: E1203 07:15:06.753146 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038\": container with ID starting with 9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038 not found: ID does not exist" containerID="9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.753189 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038"} err="failed to get container status \"9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038\": rpc error: code = NotFound desc = could not find container \"9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038\": container with ID starting with 9bfc814d1e3b04db414d870456c81bc19c9579cf6a6a53a54c6fb2dff7f0c038 not found: ID does not exist" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.759783 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.847534 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.908082 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.908148 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:06 crc kubenswrapper[4946]: E1203 07:15:06.908729 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-api" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.908743 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-api" Dec 03 07:15:06 crc kubenswrapper[4946]: E1203 07:15:06.908794 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-log" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.908802 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-log" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.909190 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-api" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.909212 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" containerName="nova-api-log" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.910657 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.910745 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.928705 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.929412 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 07:15:06 crc kubenswrapper[4946]: I1203 07:15:06.929901 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.019035 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs2xl\" (UniqueName: \"kubernetes.io/projected/6e31bf3c-7308-4e02-8f24-fe13663f4b83-kube-api-access-rs2xl\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.019131 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-config-data\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.019378 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.019540 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.019567 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e31bf3c-7308-4e02-8f24-fe13663f4b83-logs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.019613 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.122097 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.122517 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.122536 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e31bf3c-7308-4e02-8f24-fe13663f4b83-logs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.122564 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.122641 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs2xl\" (UniqueName: \"kubernetes.io/projected/6e31bf3c-7308-4e02-8f24-fe13663f4b83-kube-api-access-rs2xl\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.122674 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-config-data\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.123065 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e31bf3c-7308-4e02-8f24-fe13663f4b83-logs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.126605 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.127163 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-config-data\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.128032 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.128034 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.143037 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs2xl\" (UniqueName: \"kubernetes.io/projected/6e31bf3c-7308-4e02-8f24-fe13663f4b83-kube-api-access-rs2xl\") pod \"nova-api-0\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.256325 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.526790 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerStarted","Data":"fd75ea492b2e5d3b5ac427900e214c87df97693bea489cdea1450e4ac5fce395"} Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.527103 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerStarted","Data":"c98628a377aedb34fb875b92a9ac07a99529b16c880d351880505ad04c04f15c"} Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.547467 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.625212 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44c68706-049c-4550-ac86-3982b9d0eb70" path="/var/lib/kubelet/pods/44c68706-049c-4550-ac86-3982b9d0eb70/volumes" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.738049 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.765314 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-p2lv7"] Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.784572 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.787515 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-p2lv7"] Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.790489 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.790683 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.847055 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.847449 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mb4x\" (UniqueName: \"kubernetes.io/projected/a8760772-02a5-4231-9687-b0cd0f34f5c2-kube-api-access-4mb4x\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.847613 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-scripts\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.847789 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-config-data\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.949347 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-scripts\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.949539 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-config-data\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.949677 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.949819 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mb4x\" (UniqueName: \"kubernetes.io/projected/a8760772-02a5-4231-9687-b0cd0f34f5c2-kube-api-access-4mb4x\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.956370 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-scripts\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.956524 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.956589 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-config-data\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:07 crc kubenswrapper[4946]: I1203 07:15:07.967316 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mb4x\" (UniqueName: \"kubernetes.io/projected/a8760772-02a5-4231-9687-b0cd0f34f5c2-kube-api-access-4mb4x\") pod \"nova-cell1-cell-mapping-p2lv7\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:08 crc kubenswrapper[4946]: I1203 07:15:08.207277 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:08 crc kubenswrapper[4946]: I1203 07:15:08.548103 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerStarted","Data":"733bbc32986c6be4c2390635eddd73d705ca1fffc3e0db20cc69f0e5434d3516"} Dec 03 07:15:08 crc kubenswrapper[4946]: I1203 07:15:08.552341 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e31bf3c-7308-4e02-8f24-fe13663f4b83","Type":"ContainerStarted","Data":"ceb613b5017ed823567023db00ae90b6515a749e1c3eb970fabd70dea0c61d47"} Dec 03 07:15:08 crc kubenswrapper[4946]: I1203 07:15:08.552370 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e31bf3c-7308-4e02-8f24-fe13663f4b83","Type":"ContainerStarted","Data":"5b3607c5fa32fd1fb22ebdccb9b7f3753ee5c3d0ac62a54c6af62199aca56ed2"} Dec 03 07:15:08 crc kubenswrapper[4946]: I1203 07:15:08.552379 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e31bf3c-7308-4e02-8f24-fe13663f4b83","Type":"ContainerStarted","Data":"148da67afbfc287f4c3a7c0c12c90843c25abea27285eb35f3104de491684a48"} Dec 03 07:15:08 crc kubenswrapper[4946]: I1203 07:15:08.590633 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.590614318 podStartE2EDuration="2.590614318s" podCreationTimestamp="2025-12-03 07:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:15:08.580221828 +0000 UTC m=+1501.376911937" watchObservedRunningTime="2025-12-03 07:15:08.590614318 +0000 UTC m=+1501.387304427" Dec 03 07:15:08 crc kubenswrapper[4946]: I1203 07:15:08.690798 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-p2lv7"] Dec 03 07:15:09 crc kubenswrapper[4946]: I1203 07:15:09.568142 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerStarted","Data":"9242e69cc08af9bee9b0d9c36b79f79805314218c4453cc7949a223cbc6ad191"} Dec 03 07:15:09 crc kubenswrapper[4946]: I1203 07:15:09.570543 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p2lv7" event={"ID":"a8760772-02a5-4231-9687-b0cd0f34f5c2","Type":"ContainerStarted","Data":"39788eaaa331a03bc7d5fe3fa41f1ea2143ebafa1975eab3dca5fb0ffc7bcf07"} Dec 03 07:15:09 crc kubenswrapper[4946]: I1203 07:15:09.570620 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p2lv7" event={"ID":"a8760772-02a5-4231-9687-b0cd0f34f5c2","Type":"ContainerStarted","Data":"5d8a449956ffa9432bdafdbc4bb833af2201d27f61ff62a3a2341e1393a026fe"} Dec 03 07:15:09 crc kubenswrapper[4946]: I1203 07:15:09.606650 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-p2lv7" podStartSLOduration=2.606625182 podStartE2EDuration="2.606625182s" podCreationTimestamp="2025-12-03 07:15:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:15:09.585483684 +0000 UTC m=+1502.382173803" watchObservedRunningTime="2025-12-03 07:15:09.606625182 +0000 UTC m=+1502.403315331" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.149971 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.226283 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b9ff45c7-mpjl5"] Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.226547 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" podUID="b48f8844-fcc2-4892-826b-47e71963274f" containerName="dnsmasq-dns" containerID="cri-o://39d9643b1a86ee3d6d3b7e80844ec0028ee0f00e06200359e250cf87ad5cf169" gracePeriod=10 Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.597980 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerStarted","Data":"b88d4d8819b799ee0958eb33b8f4190198f52e90c0717a81284a3e8f801ddb08"} Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.598412 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.603162 4946 generic.go:334] "Generic (PLEG): container finished" podID="b48f8844-fcc2-4892-826b-47e71963274f" containerID="39d9643b1a86ee3d6d3b7e80844ec0028ee0f00e06200359e250cf87ad5cf169" exitCode=0 Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.604044 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" event={"ID":"b48f8844-fcc2-4892-826b-47e71963274f","Type":"ContainerDied","Data":"39d9643b1a86ee3d6d3b7e80844ec0028ee0f00e06200359e250cf87ad5cf169"} Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.697511 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.722304 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.489699883 podStartE2EDuration="5.722287725s" podCreationTimestamp="2025-12-03 07:15:05 +0000 UTC" firstStartedPulling="2025-12-03 07:15:06.621448609 +0000 UTC m=+1499.418138718" lastFinishedPulling="2025-12-03 07:15:09.854036431 +0000 UTC m=+1502.650726560" observedRunningTime="2025-12-03 07:15:10.634801403 +0000 UTC m=+1503.431491522" watchObservedRunningTime="2025-12-03 07:15:10.722287725 +0000 UTC m=+1503.518977834" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.843263 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64wz8\" (UniqueName: \"kubernetes.io/projected/b48f8844-fcc2-4892-826b-47e71963274f-kube-api-access-64wz8\") pod \"b48f8844-fcc2-4892-826b-47e71963274f\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.843382 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-swift-storage-0\") pod \"b48f8844-fcc2-4892-826b-47e71963274f\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.843444 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-svc\") pod \"b48f8844-fcc2-4892-826b-47e71963274f\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.843464 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-nb\") pod \"b48f8844-fcc2-4892-826b-47e71963274f\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.843505 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-sb\") pod \"b48f8844-fcc2-4892-826b-47e71963274f\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.843556 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-config\") pod \"b48f8844-fcc2-4892-826b-47e71963274f\" (UID: \"b48f8844-fcc2-4892-826b-47e71963274f\") " Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.859637 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48f8844-fcc2-4892-826b-47e71963274f-kube-api-access-64wz8" (OuterVolumeSpecName: "kube-api-access-64wz8") pod "b48f8844-fcc2-4892-826b-47e71963274f" (UID: "b48f8844-fcc2-4892-826b-47e71963274f"). InnerVolumeSpecName "kube-api-access-64wz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.904405 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b48f8844-fcc2-4892-826b-47e71963274f" (UID: "b48f8844-fcc2-4892-826b-47e71963274f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.908642 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b48f8844-fcc2-4892-826b-47e71963274f" (UID: "b48f8844-fcc2-4892-826b-47e71963274f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.908700 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-config" (OuterVolumeSpecName: "config") pod "b48f8844-fcc2-4892-826b-47e71963274f" (UID: "b48f8844-fcc2-4892-826b-47e71963274f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.916575 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b48f8844-fcc2-4892-826b-47e71963274f" (UID: "b48f8844-fcc2-4892-826b-47e71963274f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.917068 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b48f8844-fcc2-4892-826b-47e71963274f" (UID: "b48f8844-fcc2-4892-826b-47e71963274f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.945240 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.945276 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.945288 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.945302 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.945314 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48f8844-fcc2-4892-826b-47e71963274f-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:10 crc kubenswrapper[4946]: I1203 07:15:10.945327 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64wz8\" (UniqueName: \"kubernetes.io/projected/b48f8844-fcc2-4892-826b-47e71963274f-kube-api-access-64wz8\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:11 crc kubenswrapper[4946]: I1203 07:15:11.616656 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" event={"ID":"b48f8844-fcc2-4892-826b-47e71963274f","Type":"ContainerDied","Data":"deb8b166cac0585b69cbb4f4de2b030fdfb02b9334af953c9901ed0e45071914"} Dec 03 07:15:11 crc kubenswrapper[4946]: I1203 07:15:11.617109 4946 scope.go:117] "RemoveContainer" containerID="39d9643b1a86ee3d6d3b7e80844ec0028ee0f00e06200359e250cf87ad5cf169" Dec 03 07:15:11 crc kubenswrapper[4946]: I1203 07:15:11.616714 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b9ff45c7-mpjl5" Dec 03 07:15:11 crc kubenswrapper[4946]: I1203 07:15:11.644089 4946 scope.go:117] "RemoveContainer" containerID="76446d6826fb88bd3875eab86229dd6c5b12e905abcac321999b9ba5895aee85" Dec 03 07:15:11 crc kubenswrapper[4946]: I1203 07:15:11.653213 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b9ff45c7-mpjl5"] Dec 03 07:15:11 crc kubenswrapper[4946]: I1203 07:15:11.663933 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b9ff45c7-mpjl5"] Dec 03 07:15:13 crc kubenswrapper[4946]: I1203 07:15:13.616573 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b48f8844-fcc2-4892-826b-47e71963274f" path="/var/lib/kubelet/pods/b48f8844-fcc2-4892-826b-47e71963274f/volumes" Dec 03 07:15:13 crc kubenswrapper[4946]: I1203 07:15:13.647676 4946 generic.go:334] "Generic (PLEG): container finished" podID="a8760772-02a5-4231-9687-b0cd0f34f5c2" containerID="39788eaaa331a03bc7d5fe3fa41f1ea2143ebafa1975eab3dca5fb0ffc7bcf07" exitCode=0 Dec 03 07:15:13 crc kubenswrapper[4946]: I1203 07:15:13.647735 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p2lv7" event={"ID":"a8760772-02a5-4231-9687-b0cd0f34f5c2","Type":"ContainerDied","Data":"39788eaaa331a03bc7d5fe3fa41f1ea2143ebafa1975eab3dca5fb0ffc7bcf07"} Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.098562 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.236691 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-config-data\") pod \"a8760772-02a5-4231-9687-b0cd0f34f5c2\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.236847 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mb4x\" (UniqueName: \"kubernetes.io/projected/a8760772-02a5-4231-9687-b0cd0f34f5c2-kube-api-access-4mb4x\") pod \"a8760772-02a5-4231-9687-b0cd0f34f5c2\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.237078 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-scripts\") pod \"a8760772-02a5-4231-9687-b0cd0f34f5c2\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.237140 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-combined-ca-bundle\") pod \"a8760772-02a5-4231-9687-b0cd0f34f5c2\" (UID: \"a8760772-02a5-4231-9687-b0cd0f34f5c2\") " Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.247999 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-scripts" (OuterVolumeSpecName: "scripts") pod "a8760772-02a5-4231-9687-b0cd0f34f5c2" (UID: "a8760772-02a5-4231-9687-b0cd0f34f5c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.271694 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8760772-02a5-4231-9687-b0cd0f34f5c2-kube-api-access-4mb4x" (OuterVolumeSpecName: "kube-api-access-4mb4x") pod "a8760772-02a5-4231-9687-b0cd0f34f5c2" (UID: "a8760772-02a5-4231-9687-b0cd0f34f5c2"). InnerVolumeSpecName "kube-api-access-4mb4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.276492 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8760772-02a5-4231-9687-b0cd0f34f5c2" (UID: "a8760772-02a5-4231-9687-b0cd0f34f5c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.280698 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-config-data" (OuterVolumeSpecName: "config-data") pod "a8760772-02a5-4231-9687-b0cd0f34f5c2" (UID: "a8760772-02a5-4231-9687-b0cd0f34f5c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.339399 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.339433 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.339447 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8760772-02a5-4231-9687-b0cd0f34f5c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.339457 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mb4x\" (UniqueName: \"kubernetes.io/projected/a8760772-02a5-4231-9687-b0cd0f34f5c2-kube-api-access-4mb4x\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.676531 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p2lv7" event={"ID":"a8760772-02a5-4231-9687-b0cd0f34f5c2","Type":"ContainerDied","Data":"5d8a449956ffa9432bdafdbc4bb833af2201d27f61ff62a3a2341e1393a026fe"} Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.676591 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d8a449956ffa9432bdafdbc4bb833af2201d27f61ff62a3a2341e1393a026fe" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.676699 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p2lv7" Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.986385 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.986774 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerName="nova-api-log" containerID="cri-o://5b3607c5fa32fd1fb22ebdccb9b7f3753ee5c3d0ac62a54c6af62199aca56ed2" gracePeriod=30 Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.986980 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerName="nova-api-api" containerID="cri-o://ceb613b5017ed823567023db00ae90b6515a749e1c3eb970fabd70dea0c61d47" gracePeriod=30 Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.995416 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:15:15 crc kubenswrapper[4946]: I1203 07:15:15.995623 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="341d1e5a-8309-4cb1-8f50-2ddf556a7db0" containerName="nova-scheduler-scheduler" containerID="cri-o://4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314" gracePeriod=30 Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.029424 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.029641 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-log" containerID="cri-o://930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6" gracePeriod=30 Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.029769 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-metadata" containerID="cri-o://bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3" gracePeriod=30 Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.706918 4946 generic.go:334] "Generic (PLEG): container finished" podID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerID="930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6" exitCode=143 Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.707211 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3192ad5b-aabb-4fc7-a534-bed599cc6f1c","Type":"ContainerDied","Data":"930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6"} Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.710383 4946 generic.go:334] "Generic (PLEG): container finished" podID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerID="ceb613b5017ed823567023db00ae90b6515a749e1c3eb970fabd70dea0c61d47" exitCode=0 Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.710403 4946 generic.go:334] "Generic (PLEG): container finished" podID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerID="5b3607c5fa32fd1fb22ebdccb9b7f3753ee5c3d0ac62a54c6af62199aca56ed2" exitCode=143 Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.710418 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e31bf3c-7308-4e02-8f24-fe13663f4b83","Type":"ContainerDied","Data":"ceb613b5017ed823567023db00ae90b6515a749e1c3eb970fabd70dea0c61d47"} Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.710433 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e31bf3c-7308-4e02-8f24-fe13663f4b83","Type":"ContainerDied","Data":"5b3607c5fa32fd1fb22ebdccb9b7f3753ee5c3d0ac62a54c6af62199aca56ed2"} Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.890065 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.972834 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-config-data\") pod \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.973341 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-internal-tls-certs\") pod \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.973396 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-public-tls-certs\") pod \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.973530 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs2xl\" (UniqueName: \"kubernetes.io/projected/6e31bf3c-7308-4e02-8f24-fe13663f4b83-kube-api-access-rs2xl\") pod \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.973564 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-combined-ca-bundle\") pod \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.973610 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e31bf3c-7308-4e02-8f24-fe13663f4b83-logs\") pod \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\" (UID: \"6e31bf3c-7308-4e02-8f24-fe13663f4b83\") " Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.974460 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e31bf3c-7308-4e02-8f24-fe13663f4b83-logs" (OuterVolumeSpecName: "logs") pod "6e31bf3c-7308-4e02-8f24-fe13663f4b83" (UID: "6e31bf3c-7308-4e02-8f24-fe13663f4b83"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:16 crc kubenswrapper[4946]: I1203 07:15:16.980704 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e31bf3c-7308-4e02-8f24-fe13663f4b83-kube-api-access-rs2xl" (OuterVolumeSpecName: "kube-api-access-rs2xl") pod "6e31bf3c-7308-4e02-8f24-fe13663f4b83" (UID: "6e31bf3c-7308-4e02-8f24-fe13663f4b83"). InnerVolumeSpecName "kube-api-access-rs2xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.003383 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e31bf3c-7308-4e02-8f24-fe13663f4b83" (UID: "6e31bf3c-7308-4e02-8f24-fe13663f4b83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.028807 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6e31bf3c-7308-4e02-8f24-fe13663f4b83" (UID: "6e31bf3c-7308-4e02-8f24-fe13663f4b83"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.034652 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-config-data" (OuterVolumeSpecName: "config-data") pod "6e31bf3c-7308-4e02-8f24-fe13663f4b83" (UID: "6e31bf3c-7308-4e02-8f24-fe13663f4b83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.049299 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6e31bf3c-7308-4e02-8f24-fe13663f4b83" (UID: "6e31bf3c-7308-4e02-8f24-fe13663f4b83"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.076361 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.076405 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.076419 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs2xl\" (UniqueName: \"kubernetes.io/projected/6e31bf3c-7308-4e02-8f24-fe13663f4b83-kube-api-access-rs2xl\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.076433 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.076445 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e31bf3c-7308-4e02-8f24-fe13663f4b83-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.076457 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e31bf3c-7308-4e02-8f24-fe13663f4b83-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.524408 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.686289 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-config-data\") pod \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.687043 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmn6r\" (UniqueName: \"kubernetes.io/projected/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-kube-api-access-wmn6r\") pod \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.687106 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-combined-ca-bundle\") pod \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\" (UID: \"341d1e5a-8309-4cb1-8f50-2ddf556a7db0\") " Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.693502 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-kube-api-access-wmn6r" (OuterVolumeSpecName: "kube-api-access-wmn6r") pod "341d1e5a-8309-4cb1-8f50-2ddf556a7db0" (UID: "341d1e5a-8309-4cb1-8f50-2ddf556a7db0"). InnerVolumeSpecName "kube-api-access-wmn6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.724644 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e31bf3c-7308-4e02-8f24-fe13663f4b83","Type":"ContainerDied","Data":"148da67afbfc287f4c3a7c0c12c90843c25abea27285eb35f3104de491684a48"} Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.724712 4946 scope.go:117] "RemoveContainer" containerID="ceb613b5017ed823567023db00ae90b6515a749e1c3eb970fabd70dea0c61d47" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.726550 4946 generic.go:334] "Generic (PLEG): container finished" podID="341d1e5a-8309-4cb1-8f50-2ddf556a7db0" containerID="4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314" exitCode=0 Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.726556 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.726605 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.726637 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"341d1e5a-8309-4cb1-8f50-2ddf556a7db0","Type":"ContainerDied","Data":"4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314"} Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.726667 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"341d1e5a-8309-4cb1-8f50-2ddf556a7db0","Type":"ContainerDied","Data":"d24f3d40c3988ac37ee76502def45d5ad761af54bc70084746041b2b89f8a38a"} Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.728526 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-config-data" (OuterVolumeSpecName: "config-data") pod "341d1e5a-8309-4cb1-8f50-2ddf556a7db0" (UID: "341d1e5a-8309-4cb1-8f50-2ddf556a7db0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.734491 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "341d1e5a-8309-4cb1-8f50-2ddf556a7db0" (UID: "341d1e5a-8309-4cb1-8f50-2ddf556a7db0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.763047 4946 scope.go:117] "RemoveContainer" containerID="5b3607c5fa32fd1fb22ebdccb9b7f3753ee5c3d0ac62a54c6af62199aca56ed2" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.767341 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.791043 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmn6r\" (UniqueName: \"kubernetes.io/projected/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-kube-api-access-wmn6r\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.791077 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.791090 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341d1e5a-8309-4cb1-8f50-2ddf556a7db0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.791676 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.796821 4946 scope.go:117] "RemoveContainer" containerID="4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.802750 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:17 crc kubenswrapper[4946]: E1203 07:15:17.803460 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8760772-02a5-4231-9687-b0cd0f34f5c2" containerName="nova-manage" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803482 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8760772-02a5-4231-9687-b0cd0f34f5c2" containerName="nova-manage" Dec 03 07:15:17 crc kubenswrapper[4946]: E1203 07:15:17.803503 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerName="nova-api-api" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803512 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerName="nova-api-api" Dec 03 07:15:17 crc kubenswrapper[4946]: E1203 07:15:17.803525 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341d1e5a-8309-4cb1-8f50-2ddf556a7db0" containerName="nova-scheduler-scheduler" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803534 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="341d1e5a-8309-4cb1-8f50-2ddf556a7db0" containerName="nova-scheduler-scheduler" Dec 03 07:15:17 crc kubenswrapper[4946]: E1203 07:15:17.803550 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48f8844-fcc2-4892-826b-47e71963274f" containerName="init" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803558 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48f8844-fcc2-4892-826b-47e71963274f" containerName="init" Dec 03 07:15:17 crc kubenswrapper[4946]: E1203 07:15:17.803579 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerName="nova-api-log" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803587 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerName="nova-api-log" Dec 03 07:15:17 crc kubenswrapper[4946]: E1203 07:15:17.803602 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48f8844-fcc2-4892-826b-47e71963274f" containerName="dnsmasq-dns" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803609 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48f8844-fcc2-4892-826b-47e71963274f" containerName="dnsmasq-dns" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803830 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerName="nova-api-api" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803862 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="341d1e5a-8309-4cb1-8f50-2ddf556a7db0" containerName="nova-scheduler-scheduler" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803873 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8760772-02a5-4231-9687-b0cd0f34f5c2" containerName="nova-manage" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803892 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48f8844-fcc2-4892-826b-47e71963274f" containerName="dnsmasq-dns" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.803901 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" containerName="nova-api-log" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.806354 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.808717 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.808776 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.811317 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.821041 4946 scope.go:117] "RemoveContainer" containerID="4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314" Dec 03 07:15:17 crc kubenswrapper[4946]: E1203 07:15:17.821501 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314\": container with ID starting with 4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314 not found: ID does not exist" containerID="4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.821541 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314"} err="failed to get container status \"4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314\": rpc error: code = NotFound desc = could not find container \"4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314\": container with ID starting with 4faef96549f3f7d55c4911bf456ea42b8e58da8a6cf7e013f7321d33fe068314 not found: ID does not exist" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.822106 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.892788 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-logs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.892892 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.892951 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvnbh\" (UniqueName: \"kubernetes.io/projected/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-kube-api-access-zvnbh\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.893147 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-config-data\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.893227 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-public-tls-certs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.893288 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.995185 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-config-data\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.995246 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-public-tls-certs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.995279 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.995369 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-logs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.995417 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.995441 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvnbh\" (UniqueName: \"kubernetes.io/projected/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-kube-api-access-zvnbh\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.996041 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-logs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:17 crc kubenswrapper[4946]: I1203 07:15:17.999504 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.001529 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-config-data\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.002340 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-public-tls-certs\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.003763 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.015774 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvnbh\" (UniqueName: \"kubernetes.io/projected/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-kube-api-access-zvnbh\") pod \"nova-api-0\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " pod="openstack/nova-api-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.074433 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.087459 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.105666 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.106929 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.107107 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.109252 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.140953 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.199673 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwmjt\" (UniqueName: \"kubernetes.io/projected/e5415d2e-cf73-4dd8-900d-211b957e829f-kube-api-access-wwmjt\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.199729 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-config-data\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.199783 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.303716 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwmjt\" (UniqueName: \"kubernetes.io/projected/e5415d2e-cf73-4dd8-900d-211b957e829f-kube-api-access-wwmjt\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.303797 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-config-data\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.303852 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.308360 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-config-data\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.315498 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.322401 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwmjt\" (UniqueName: \"kubernetes.io/projected/e5415d2e-cf73-4dd8-900d-211b957e829f-kube-api-access-wwmjt\") pod \"nova-scheduler-0\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.550323 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.592643 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:15:18 crc kubenswrapper[4946]: W1203 07:15:18.596405 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4dc4eff7_f011_4265_b9dd_4e6bcdcb08df.slice/crio-47ddc59169e2f57c7bc4a899a86665600d584abf00eecf20b360926ed8b20fb4 WatchSource:0}: Error finding container 47ddc59169e2f57c7bc4a899a86665600d584abf00eecf20b360926ed8b20fb4: Status 404 returned error can't find the container with id 47ddc59169e2f57c7bc4a899a86665600d584abf00eecf20b360926ed8b20fb4 Dec 03 07:15:18 crc kubenswrapper[4946]: I1203 07:15:18.737851 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df","Type":"ContainerStarted","Data":"47ddc59169e2f57c7bc4a899a86665600d584abf00eecf20b360926ed8b20fb4"} Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.094728 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.189093 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:57772->10.217.0.191:8775: read: connection reset by peer" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.190296 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:57780->10.217.0.191:8775: read: connection reset by peer" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.604406 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="341d1e5a-8309-4cb1-8f50-2ddf556a7db0" path="/var/lib/kubelet/pods/341d1e5a-8309-4cb1-8f50-2ddf556a7db0/volumes" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.605709 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e31bf3c-7308-4e02-8f24-fe13663f4b83" path="/var/lib/kubelet/pods/6e31bf3c-7308-4e02-8f24-fe13663f4b83/volumes" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.639023 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.735530 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-config-data\") pod \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.735671 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-logs\") pod \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.735793 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-combined-ca-bundle\") pod \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.735881 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-nova-metadata-tls-certs\") pod \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.735953 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlng5\" (UniqueName: \"kubernetes.io/projected/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-kube-api-access-qlng5\") pod \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\" (UID: \"3192ad5b-aabb-4fc7-a534-bed599cc6f1c\") " Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.736634 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-logs" (OuterVolumeSpecName: "logs") pod "3192ad5b-aabb-4fc7-a534-bed599cc6f1c" (UID: "3192ad5b-aabb-4fc7-a534-bed599cc6f1c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.741855 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-kube-api-access-qlng5" (OuterVolumeSpecName: "kube-api-access-qlng5") pod "3192ad5b-aabb-4fc7-a534-bed599cc6f1c" (UID: "3192ad5b-aabb-4fc7-a534-bed599cc6f1c"). InnerVolumeSpecName "kube-api-access-qlng5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.759075 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df","Type":"ContainerStarted","Data":"434d3710c0181e612e1826bfbdad35d34bd77749c41c0c2a77720e391d7d862e"} Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.759125 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df","Type":"ContainerStarted","Data":"6c6cec4e34e14e36e892d11434e257757b6502d3fb6f96b9a2e74b1f7a78898a"} Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.762699 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5415d2e-cf73-4dd8-900d-211b957e829f","Type":"ContainerStarted","Data":"5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6"} Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.762771 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5415d2e-cf73-4dd8-900d-211b957e829f","Type":"ContainerStarted","Data":"db27503cb03cb197511a3037f23064a2773acab8bcb40b49a1d67ee7d47f7e26"} Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.766280 4946 generic.go:334] "Generic (PLEG): container finished" podID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerID="bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3" exitCode=0 Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.766320 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3192ad5b-aabb-4fc7-a534-bed599cc6f1c","Type":"ContainerDied","Data":"bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3"} Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.766349 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3192ad5b-aabb-4fc7-a534-bed599cc6f1c","Type":"ContainerDied","Data":"36302fee7c7aa6579d11f5f9eb6797028eeacbf62be2d66222dbbf9af202d055"} Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.766370 4946 scope.go:117] "RemoveContainer" containerID="bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.766542 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.779912 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3192ad5b-aabb-4fc7-a534-bed599cc6f1c" (UID: "3192ad5b-aabb-4fc7-a534-bed599cc6f1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.782288 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-config-data" (OuterVolumeSpecName: "config-data") pod "3192ad5b-aabb-4fc7-a534-bed599cc6f1c" (UID: "3192ad5b-aabb-4fc7-a534-bed599cc6f1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.793175 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.793158404 podStartE2EDuration="2.793158404s" podCreationTimestamp="2025-12-03 07:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:15:19.791107879 +0000 UTC m=+1512.587797988" watchObservedRunningTime="2025-12-03 07:15:19.793158404 +0000 UTC m=+1512.589848513" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.805388 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3192ad5b-aabb-4fc7-a534-bed599cc6f1c" (UID: "3192ad5b-aabb-4fc7-a534-bed599cc6f1c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.810028 4946 scope.go:117] "RemoveContainer" containerID="930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.819096 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.819077021 podStartE2EDuration="1.819077021s" podCreationTimestamp="2025-12-03 07:15:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:15:19.812456673 +0000 UTC m=+1512.609146782" watchObservedRunningTime="2025-12-03 07:15:19.819077021 +0000 UTC m=+1512.615767150" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.835594 4946 scope.go:117] "RemoveContainer" containerID="bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3" Dec 03 07:15:19 crc kubenswrapper[4946]: E1203 07:15:19.835995 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3\": container with ID starting with bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3 not found: ID does not exist" containerID="bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.836022 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3"} err="failed to get container status \"bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3\": rpc error: code = NotFound desc = could not find container \"bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3\": container with ID starting with bdb5b61c67950e1c04fffb4ba120b6f83b3d183319ee368c36bbe265d8b6cde3 not found: ID does not exist" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.836042 4946 scope.go:117] "RemoveContainer" containerID="930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6" Dec 03 07:15:19 crc kubenswrapper[4946]: E1203 07:15:19.836250 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6\": container with ID starting with 930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6 not found: ID does not exist" containerID="930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.836280 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6"} err="failed to get container status \"930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6\": rpc error: code = NotFound desc = could not find container \"930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6\": container with ID starting with 930da10231234b1bab9390e2f4fabbcdda7c7e6a803cf46559383604bd649dd6 not found: ID does not exist" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.838498 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.838523 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.838650 4946 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.838930 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlng5\" (UniqueName: \"kubernetes.io/projected/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-kube-api-access-qlng5\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:19 crc kubenswrapper[4946]: I1203 07:15:19.838956 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3192ad5b-aabb-4fc7-a534-bed599cc6f1c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.102244 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.110725 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.134595 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:15:20 crc kubenswrapper[4946]: E1203 07:15:20.135149 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-log" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.135182 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-log" Dec 03 07:15:20 crc kubenswrapper[4946]: E1203 07:15:20.135208 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-metadata" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.135219 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-metadata" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.135537 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-metadata" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.135594 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" containerName="nova-metadata-log" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.137093 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.139779 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.140059 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.176727 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.264566 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/192c2964-d6c9-404e-8cca-cac01c750c5b-logs\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.264654 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.264681 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-config-data\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.264730 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp7fn\" (UniqueName: \"kubernetes.io/projected/192c2964-d6c9-404e-8cca-cac01c750c5b-kube-api-access-qp7fn\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.264839 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.366425 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/192c2964-d6c9-404e-8cca-cac01c750c5b-logs\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.366495 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.366512 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-config-data\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.366552 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp7fn\" (UniqueName: \"kubernetes.io/projected/192c2964-d6c9-404e-8cca-cac01c750c5b-kube-api-access-qp7fn\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.366621 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.367729 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/192c2964-d6c9-404e-8cca-cac01c750c5b-logs\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.371426 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.371842 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.372194 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-config-data\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.387248 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp7fn\" (UniqueName: \"kubernetes.io/projected/192c2964-d6c9-404e-8cca-cac01c750c5b-kube-api-access-qp7fn\") pod \"nova-metadata-0\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " pod="openstack/nova-metadata-0" Dec 03 07:15:20 crc kubenswrapper[4946]: I1203 07:15:20.524018 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:15:21 crc kubenswrapper[4946]: W1203 07:15:21.051907 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod192c2964_d6c9_404e_8cca_cac01c750c5b.slice/crio-f052ad62e4e6729b8aeeb5ac686c3fd61a7a04841e36e2178b8a4744bbe0b9fe WatchSource:0}: Error finding container f052ad62e4e6729b8aeeb5ac686c3fd61a7a04841e36e2178b8a4744bbe0b9fe: Status 404 returned error can't find the container with id f052ad62e4e6729b8aeeb5ac686c3fd61a7a04841e36e2178b8a4744bbe0b9fe Dec 03 07:15:21 crc kubenswrapper[4946]: I1203 07:15:21.058081 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:15:21 crc kubenswrapper[4946]: I1203 07:15:21.609832 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3192ad5b-aabb-4fc7-a534-bed599cc6f1c" path="/var/lib/kubelet/pods/3192ad5b-aabb-4fc7-a534-bed599cc6f1c/volumes" Dec 03 07:15:21 crc kubenswrapper[4946]: I1203 07:15:21.794656 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"192c2964-d6c9-404e-8cca-cac01c750c5b","Type":"ContainerStarted","Data":"99eb604fda1c1f8beeca26feebc6e7d241d1c8fcbc5022252d4d91c35a9e66a0"} Dec 03 07:15:21 crc kubenswrapper[4946]: I1203 07:15:21.795046 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"192c2964-d6c9-404e-8cca-cac01c750c5b","Type":"ContainerStarted","Data":"04a7a8335ee4dd2cc59886e31042cdc9337381a24ca52a361f5d5f35b6bdead5"} Dec 03 07:15:21 crc kubenswrapper[4946]: I1203 07:15:21.795069 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"192c2964-d6c9-404e-8cca-cac01c750c5b","Type":"ContainerStarted","Data":"f052ad62e4e6729b8aeeb5ac686c3fd61a7a04841e36e2178b8a4744bbe0b9fe"} Dec 03 07:15:21 crc kubenswrapper[4946]: I1203 07:15:21.822012 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.821992327 podStartE2EDuration="1.821992327s" podCreationTimestamp="2025-12-03 07:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:15:21.81912836 +0000 UTC m=+1514.615818479" watchObservedRunningTime="2025-12-03 07:15:21.821992327 +0000 UTC m=+1514.618682436" Dec 03 07:15:23 crc kubenswrapper[4946]: I1203 07:15:23.039644 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:15:23 crc kubenswrapper[4946]: I1203 07:15:23.040180 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:15:23 crc kubenswrapper[4946]: I1203 07:15:23.551341 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 07:15:25 crc kubenswrapper[4946]: I1203 07:15:25.525162 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 07:15:25 crc kubenswrapper[4946]: I1203 07:15:25.525547 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 07:15:28 crc kubenswrapper[4946]: I1203 07:15:28.142017 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 07:15:28 crc kubenswrapper[4946]: I1203 07:15:28.143990 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 07:15:28 crc kubenswrapper[4946]: I1203 07:15:28.551083 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 07:15:28 crc kubenswrapper[4946]: I1203 07:15:28.597194 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 07:15:28 crc kubenswrapper[4946]: I1203 07:15:28.936444 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 07:15:29 crc kubenswrapper[4946]: I1203 07:15:29.163900 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 07:15:29 crc kubenswrapper[4946]: I1203 07:15:29.163986 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 07:15:30 crc kubenswrapper[4946]: I1203 07:15:30.524557 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 07:15:30 crc kubenswrapper[4946]: I1203 07:15:30.525891 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 07:15:31 crc kubenswrapper[4946]: I1203 07:15:31.535927 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 07:15:31 crc kubenswrapper[4946]: I1203 07:15:31.535970 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 07:15:36 crc kubenswrapper[4946]: I1203 07:15:36.011122 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 07:15:38 crc kubenswrapper[4946]: I1203 07:15:38.149731 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 07:15:38 crc kubenswrapper[4946]: I1203 07:15:38.150422 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 07:15:38 crc kubenswrapper[4946]: I1203 07:15:38.150791 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 07:15:38 crc kubenswrapper[4946]: I1203 07:15:38.150827 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 07:15:38 crc kubenswrapper[4946]: I1203 07:15:38.157393 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 07:15:38 crc kubenswrapper[4946]: I1203 07:15:38.158261 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 07:15:40 crc kubenswrapper[4946]: I1203 07:15:40.533393 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 07:15:40 crc kubenswrapper[4946]: I1203 07:15:40.534308 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 07:15:40 crc kubenswrapper[4946]: I1203 07:15:40.539790 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 07:15:41 crc kubenswrapper[4946]: I1203 07:15:41.019202 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.118094 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xcjbh"] Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.121170 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.137234 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xcjbh"] Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.180727 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-catalog-content\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.180885 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8czkh\" (UniqueName: \"kubernetes.io/projected/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-kube-api-access-8czkh\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.181201 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-utilities\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.282709 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8czkh\" (UniqueName: \"kubernetes.io/projected/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-kube-api-access-8czkh\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.282861 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-utilities\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.282959 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-catalog-content\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.283333 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-utilities\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.283367 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-catalog-content\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.308044 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8czkh\" (UniqueName: \"kubernetes.io/projected/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-kube-api-access-8czkh\") pod \"redhat-operators-xcjbh\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.454452 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:43 crc kubenswrapper[4946]: I1203 07:15:43.900785 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xcjbh"] Dec 03 07:15:44 crc kubenswrapper[4946]: I1203 07:15:44.035239 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjbh" event={"ID":"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb","Type":"ContainerStarted","Data":"69e66f0ed3549111e29aa3b76ae5254f6c3de0f690272416b58064d884a00d41"} Dec 03 07:15:45 crc kubenswrapper[4946]: I1203 07:15:45.052381 4946 generic.go:334] "Generic (PLEG): container finished" podID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerID="6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7" exitCode=0 Dec 03 07:15:45 crc kubenswrapper[4946]: I1203 07:15:45.052465 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjbh" event={"ID":"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb","Type":"ContainerDied","Data":"6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7"} Dec 03 07:15:46 crc kubenswrapper[4946]: I1203 07:15:46.069934 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjbh" event={"ID":"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb","Type":"ContainerStarted","Data":"3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee"} Dec 03 07:15:48 crc kubenswrapper[4946]: I1203 07:15:48.095562 4946 generic.go:334] "Generic (PLEG): container finished" podID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerID="3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee" exitCode=0 Dec 03 07:15:48 crc kubenswrapper[4946]: I1203 07:15:48.095634 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjbh" event={"ID":"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb","Type":"ContainerDied","Data":"3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee"} Dec 03 07:15:50 crc kubenswrapper[4946]: I1203 07:15:50.132570 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjbh" event={"ID":"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb","Type":"ContainerStarted","Data":"d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e"} Dec 03 07:15:50 crc kubenswrapper[4946]: I1203 07:15:50.151463 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xcjbh" podStartSLOduration=2.4870995909999998 podStartE2EDuration="7.151445992s" podCreationTimestamp="2025-12-03 07:15:43 +0000 UTC" firstStartedPulling="2025-12-03 07:15:45.056013416 +0000 UTC m=+1537.852703575" lastFinishedPulling="2025-12-03 07:15:49.720359857 +0000 UTC m=+1542.517049976" observedRunningTime="2025-12-03 07:15:50.149984443 +0000 UTC m=+1542.946674552" watchObservedRunningTime="2025-12-03 07:15:50.151445992 +0000 UTC m=+1542.948136111" Dec 03 07:15:53 crc kubenswrapper[4946]: I1203 07:15:53.040243 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:15:53 crc kubenswrapper[4946]: I1203 07:15:53.041090 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:15:53 crc kubenswrapper[4946]: I1203 07:15:53.455595 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:53 crc kubenswrapper[4946]: I1203 07:15:53.455697 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:15:54 crc kubenswrapper[4946]: I1203 07:15:54.539328 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xcjbh" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="registry-server" probeResult="failure" output=< Dec 03 07:15:54 crc kubenswrapper[4946]: timeout: failed to connect service ":50051" within 1s Dec 03 07:15:54 crc kubenswrapper[4946]: > Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.107019 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.107656 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerName="cinder-scheduler" containerID="cri-o://fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219" gracePeriod=30 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.108076 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerName="probe" containerID="cri-o://ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374" gracePeriod=30 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.171230 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.171454 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="788c3bf1-3291-4324-8a4c-dc06f1027f5b" containerName="openstackclient" containerID="cri-o://4aaa4e05b46fa6c68df1bd9487eadf3885468fbce3baff034433d3bc2c281e0b" gracePeriod=2 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.197947 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.316169 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.316402 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api-log" containerID="cri-o://0c1e1e5afa702c1d51ea8d3cbdd531ef58625215a8edb86478cd68e3ae7a611f" gracePeriod=30 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.316506 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api" containerID="cri-o://6d8145ae8f7303d9ebe9697e32c04cdfb3a46dd8ce1dc51d2f81b3c03b54e2c8" gracePeriod=30 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.350778 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.408992 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancec521-account-delete-d62ms"] Dec 03 07:16:03 crc kubenswrapper[4946]: E1203 07:16:03.409375 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788c3bf1-3291-4324-8a4c-dc06f1027f5b" containerName="openstackclient" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.409393 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="788c3bf1-3291-4324-8a4c-dc06f1027f5b" containerName="openstackclient" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.409595 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="788c3bf1-3291-4324-8a4c-dc06f1027f5b" containerName="openstackclient" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.410203 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.424380 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancec521-account-delete-d62ms"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.472249 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron6e16-account-delete-snvqr"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.473924 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.532261 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.532921 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerName="openstack-network-exporter" containerID="cri-o://dbf1d62862ef8e193ae4fb4237e071cfc9defb2902e956f02581df96ce0e9858" gracePeriod=300 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.547886 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e57ed09-f7ce-403e-9cb5-676ae1c64117-operator-scripts\") pod \"glancec521-account-delete-d62ms\" (UID: \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\") " pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.548106 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g647z\" (UniqueName: \"kubernetes.io/projected/0e57ed09-f7ce-403e-9cb5-676ae1c64117-kube-api-access-g647z\") pod \"glancec521-account-delete-d62ms\" (UID: \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\") " pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:03 crc kubenswrapper[4946]: E1203 07:16:03.549311 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 07:16:03 crc kubenswrapper[4946]: E1203 07:16:03.549368 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data podName:854575dc-d77d-43fe-a7ff-495ef76ab456 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:04.049353826 +0000 UTC m=+1556.846043935 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data") pod "rabbitmq-server-0" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456") : configmap "rabbitmq-config-data" not found Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.561137 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron6e16-account-delete-snvqr"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.594022 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.639590 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.639894 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-9q4dr"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.640339 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="8260f89c-239d-42e6-b187-c912013d5eca" containerName="openstack-network-exporter" containerID="cri-o://bbdd199402f5e0e7ad481bf616fe43d3edcdb9ecdbe811804101a2daa767e736" gracePeriod=300 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.649148 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-9q4dr"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.652298 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e57ed09-f7ce-403e-9cb5-676ae1c64117-operator-scripts\") pod \"glancec521-account-delete-d62ms\" (UID: \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\") " pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.652347 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts\") pod \"neutron6e16-account-delete-snvqr\" (UID: \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\") " pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.652479 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wjw8\" (UniqueName: \"kubernetes.io/projected/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-kube-api-access-7wjw8\") pod \"neutron6e16-account-delete-snvqr\" (UID: \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\") " pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.652525 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g647z\" (UniqueName: \"kubernetes.io/projected/0e57ed09-f7ce-403e-9cb5-676ae1c64117-kube-api-access-g647z\") pod \"glancec521-account-delete-d62ms\" (UID: \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\") " pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.653469 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e57ed09-f7ce-403e-9cb5-676ae1c64117-operator-scripts\") pod \"glancec521-account-delete-d62ms\" (UID: \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\") " pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.677524 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-qzv4q"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.688316 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g647z\" (UniqueName: \"kubernetes.io/projected/0e57ed09-f7ce-403e-9cb5-676ae1c64117-kube-api-access-g647z\") pod \"glancec521-account-delete-d62ms\" (UID: \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\") " pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.696705 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement0960-account-delete-pc5x5"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.697928 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.700881 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerName="ovsdbserver-sb" containerID="cri-o://306f0aa54817d7e3ac475556197a16b148ab3536c5c0e1a6f77075f7234d7b98" gracePeriod=300 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.727094 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-qzv4q"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.736874 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.743456 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement0960-account-delete-pc5x5"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.754062 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts\") pod \"neutron6e16-account-delete-snvqr\" (UID: \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\") " pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.754270 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wjw8\" (UniqueName: \"kubernetes.io/projected/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-kube-api-access-7wjw8\") pod \"neutron6e16-account-delete-snvqr\" (UID: \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\") " pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.755590 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.771393 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts\") pod \"neutron6e16-account-delete-snvqr\" (UID: \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\") " pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.834720 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicaneae6-account-delete-l9467"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.836085 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.840919 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wjw8\" (UniqueName: \"kubernetes.io/projected/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-kube-api-access-7wjw8\") pod \"neutron6e16-account-delete-snvqr\" (UID: \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\") " pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.858764 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8pc7\" (UniqueName: \"kubernetes.io/projected/572a50eb-42e9-4897-9925-4073b5f6a35d-kube-api-access-c8pc7\") pod \"placement0960-account-delete-pc5x5\" (UID: \"572a50eb-42e9-4897-9925-4073b5f6a35d\") " pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.858930 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts\") pod \"placement0960-account-delete-pc5x5\" (UID: \"572a50eb-42e9-4897-9925-4073b5f6a35d\") " pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.870153 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicaneae6-account-delete-l9467"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.886299 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="8260f89c-239d-42e6-b187-c912013d5eca" containerName="ovsdbserver-nb" containerID="cri-o://5b78ba643d512f9707208df0f4f8fcbbe76854d669c9d51da5ef8396964bb755" gracePeriod=300 Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.940909 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder5125-account-delete-lrpcc"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.942682 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.951838 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder5125-account-delete-lrpcc"] Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.962312 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ttpl\" (UniqueName: \"kubernetes.io/projected/85a6dcc4-23cc-4f57-8714-06d0ef25c680-kube-api-access-2ttpl\") pod \"barbicaneae6-account-delete-l9467\" (UID: \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\") " pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.963366 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts\") pod \"placement0960-account-delete-pc5x5\" (UID: \"572a50eb-42e9-4897-9925-4073b5f6a35d\") " pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.964398 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts\") pod \"placement0960-account-delete-pc5x5\" (UID: \"572a50eb-42e9-4897-9925-4073b5f6a35d\") " pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.964444 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts\") pod \"barbicaneae6-account-delete-l9467\" (UID: \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\") " pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:03 crc kubenswrapper[4946]: I1203 07:16:03.964608 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8pc7\" (UniqueName: \"kubernetes.io/projected/572a50eb-42e9-4897-9925-4073b5f6a35d-kube-api-access-c8pc7\") pod \"placement0960-account-delete-pc5x5\" (UID: \"572a50eb-42e9-4897-9925-4073b5f6a35d\") " pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.020798 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.021314 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="ovn-northd" containerID="cri-o://1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.022944 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="openstack-network-exporter" containerID="cri-o://6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.062490 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8pc7\" (UniqueName: \"kubernetes.io/projected/572a50eb-42e9-4897-9925-4073b5f6a35d-kube-api-access-c8pc7\") pod \"placement0960-account-delete-pc5x5\" (UID: \"572a50eb-42e9-4897-9925-4073b5f6a35d\") " pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.069564 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts\") pod \"barbicaneae6-account-delete-l9467\" (UID: \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\") " pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.070369 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts\") pod \"barbicaneae6-account-delete-l9467\" (UID: \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\") " pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:04 crc kubenswrapper[4946]: E1203 07:16:04.071544 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.069718 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf286a1-0b76-4e66-817c-7d7ce84edaa2-operator-scripts\") pod \"cinder5125-account-delete-lrpcc\" (UID: \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\") " pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:04 crc kubenswrapper[4946]: E1203 07:16:04.071628 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data podName:854575dc-d77d-43fe-a7ff-495ef76ab456 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:05.071608911 +0000 UTC m=+1557.868299020 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data") pod "rabbitmq-server-0" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456") : configmap "rabbitmq-config-data" not found Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.071847 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ttpl\" (UniqueName: \"kubernetes.io/projected/85a6dcc4-23cc-4f57-8714-06d0ef25c680-kube-api-access-2ttpl\") pod \"barbicaneae6-account-delete-l9467\" (UID: \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\") " pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.071905 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46dbr\" (UniqueName: \"kubernetes.io/projected/baf286a1-0b76-4e66-817c-7d7ce84edaa2-kube-api-access-46dbr\") pod \"cinder5125-account-delete-lrpcc\" (UID: \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\") " pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.107160 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.111020 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xcjbh"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.124921 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.146644 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ttpl\" (UniqueName: \"kubernetes.io/projected/85a6dcc4-23cc-4f57-8714-06d0ef25c680-kube-api-access-2ttpl\") pod \"barbicaneae6-account-delete-l9467\" (UID: \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\") " pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.178191 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf286a1-0b76-4e66-817c-7d7ce84edaa2-operator-scripts\") pod \"cinder5125-account-delete-lrpcc\" (UID: \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\") " pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.178298 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46dbr\" (UniqueName: \"kubernetes.io/projected/baf286a1-0b76-4e66-817c-7d7ce84edaa2-kube-api-access-46dbr\") pod \"cinder5125-account-delete-lrpcc\" (UID: \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\") " pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.179421 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf286a1-0b76-4e66-817c-7d7ce84edaa2-operator-scripts\") pod \"cinder5125-account-delete-lrpcc\" (UID: \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\") " pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.203684 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46dbr\" (UniqueName: \"kubernetes.io/projected/baf286a1-0b76-4e66-817c-7d7ce84edaa2-kube-api-access-46dbr\") pod \"cinder5125-account-delete-lrpcc\" (UID: \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\") " pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.225477 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.250617 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi2e82-account-delete-77lgb"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.253895 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.289955 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qhs9v"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.315189 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qhs9v"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.316541 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.317152 4946 generic.go:334] "Generic (PLEG): container finished" podID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerID="0c1e1e5afa702c1d51ea8d3cbdd531ef58625215a8edb86478cd68e3ae7a611f" exitCode=143 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.317214 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a3521a3-0f30-4afe-b5f8-efeb6f838e72","Type":"ContainerDied","Data":"0c1e1e5afa702c1d51ea8d3cbdd531ef58625215a8edb86478cd68e3ae7a611f"} Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.327553 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7b3d3641-40ca-4391-b7f2-811a85e9c99e/ovsdbserver-sb/0.log" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.327785 4946 generic.go:334] "Generic (PLEG): container finished" podID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerID="dbf1d62862ef8e193ae4fb4237e071cfc9defb2902e956f02581df96ce0e9858" exitCode=2 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.327859 4946 generic.go:334] "Generic (PLEG): container finished" podID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerID="306f0aa54817d7e3ac475556197a16b148ab3536c5c0e1a6f77075f7234d7b98" exitCode=143 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.327972 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b3d3641-40ca-4391-b7f2-811a85e9c99e","Type":"ContainerDied","Data":"dbf1d62862ef8e193ae4fb4237e071cfc9defb2902e956f02581df96ce0e9858"} Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.328058 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b3d3641-40ca-4391-b7f2-811a85e9c99e","Type":"ContainerDied","Data":"306f0aa54817d7e3ac475556197a16b148ab3536c5c0e1a6f77075f7234d7b98"} Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.331220 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8260f89c-239d-42e6-b187-c912013d5eca/ovsdbserver-nb/0.log" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.331261 4946 generic.go:334] "Generic (PLEG): container finished" podID="8260f89c-239d-42e6-b187-c912013d5eca" containerID="bbdd199402f5e0e7ad481bf616fe43d3edcdb9ecdbe811804101a2daa767e736" exitCode=2 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.331280 4946 generic.go:334] "Generic (PLEG): container finished" podID="8260f89c-239d-42e6-b187-c912013d5eca" containerID="5b78ba643d512f9707208df0f4f8fcbbe76854d669c9d51da5ef8396964bb755" exitCode=143 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.331346 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8260f89c-239d-42e6-b187-c912013d5eca","Type":"ContainerDied","Data":"bbdd199402f5e0e7ad481bf616fe43d3edcdb9ecdbe811804101a2daa767e736"} Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.331377 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8260f89c-239d-42e6-b187-c912013d5eca","Type":"ContainerDied","Data":"5b78ba643d512f9707208df0f4f8fcbbe76854d669c9d51da5ef8396964bb755"} Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.341889 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi2e82-account-delete-77lgb"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.354845 4946 generic.go:334] "Generic (PLEG): container finished" podID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerID="6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82" exitCode=2 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.355265 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"80557a83-debb-4e3a-a2d0-bb5ac72d824c","Type":"ContainerDied","Data":"6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82"} Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.378658 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0a5c7-account-delete-j2fhk"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.384076 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.395332 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5gxj\" (UniqueName: \"kubernetes.io/projected/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-kube-api-access-c5gxj\") pod \"novaapi2e82-account-delete-77lgb\" (UID: \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\") " pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.395596 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts\") pod \"novaapi2e82-account-delete-77lgb\" (UID: \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\") " pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.405784 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0a5c7-account-delete-j2fhk"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.422517 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.439939 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-qnf4v"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.454937 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-qnf4v"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.484414 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-9twd7"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.497035 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbpfg\" (UniqueName: \"kubernetes.io/projected/0bc0555a-6c39-4e57-8c30-6a6fc422376a-kube-api-access-nbpfg\") pod \"novacell0a5c7-account-delete-j2fhk\" (UID: \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\") " pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.497081 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts\") pod \"novacell0a5c7-account-delete-j2fhk\" (UID: \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\") " pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.497142 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5gxj\" (UniqueName: \"kubernetes.io/projected/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-kube-api-access-c5gxj\") pod \"novaapi2e82-account-delete-77lgb\" (UID: \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\") " pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.500860 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts\") pod \"novaapi2e82-account-delete-77lgb\" (UID: \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\") " pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.509673 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts\") pod \"novaapi2e82-account-delete-77lgb\" (UID: \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\") " pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.536833 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5gxj\" (UniqueName: \"kubernetes.io/projected/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-kube-api-access-c5gxj\") pod \"novaapi2e82-account-delete-77lgb\" (UID: \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\") " pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.572424 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-9twd7"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.584991 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-fh88p"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.585194 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-fh88p" podUID="df3b0afc-1419-44ce-89f7-b1aca189e145" containerName="openstack-network-exporter" containerID="cri-o://2eb06361aba92c8a4af213131c660c362c2c6949de5340c6fe246ad8413ed5fd" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.594252 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-sk56f"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.607239 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbpfg\" (UniqueName: \"kubernetes.io/projected/0bc0555a-6c39-4e57-8c30-6a6fc422376a-kube-api-access-nbpfg\") pod \"novacell0a5c7-account-delete-j2fhk\" (UID: \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\") " pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.607309 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts\") pod \"novacell0a5c7-account-delete-j2fhk\" (UID: \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\") " pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.608091 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts\") pod \"novacell0a5c7-account-delete-j2fhk\" (UID: \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\") " pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:04 crc kubenswrapper[4946]: E1203 07:16:04.610829 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 07:16:04 crc kubenswrapper[4946]: E1203 07:16:04.610883 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data podName:2f821218-d2a4-4f97-a924-199c056c7c55 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:05.110871223 +0000 UTC m=+1557.907561332 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data") pod "rabbitmq-cell1-server-0" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55") : configmap "rabbitmq-cell1-config-data" not found Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.613927 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.624843 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hc87l"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.639647 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbpfg\" (UniqueName: \"kubernetes.io/projected/0bc0555a-6c39-4e57-8c30-6a6fc422376a-kube-api-access-nbpfg\") pod \"novacell0a5c7-account-delete-j2fhk\" (UID: \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\") " pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.640611 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86855dfc4c-jdcjc"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.640868 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-86855dfc4c-jdcjc" podUID="cc385f93-2111-42cf-93be-d4cc82225570" containerName="neutron-api" containerID="cri-o://a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.640984 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-86855dfc4c-jdcjc" podUID="cc385f93-2111-42cf-93be-d4cc82225570" containerName="neutron-httpd" containerID="cri-o://01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.661523 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.668679 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerName="glance-log" containerID="cri-o://33b47ed2f51fb1905563dad28b7eecdde497b32b4bd342ccb3247ecc1b377e2f" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.668866 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerName="glance-httpd" containerID="cri-o://7289e3069aa5ae14b8fa7cb3805ddb5abba4f40160f9caf5ebdf388dff4bf65a" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.708652 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-s624s"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.724284 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.738827 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-s624s"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.763987 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-p2lv7"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.778614 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-p2lv7"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.791577 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7b3d3641-40ca-4391-b7f2-811a85e9c99e/ovsdbserver-sb/0.log" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.791649 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.792008 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-mdtz2"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.804099 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-mdtz2"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.837754 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.837982 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerName="glance-log" containerID="cri-o://86788e20897ae74696026dd8776710b0bda123a3f5dc62f4eebb35926ddd9811" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.838110 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerName="glance-httpd" containerID="cri-o://6c046b12932defb2b1b37ab4cba5b712c6fb5e6d71e6a3ef814f1586ed8b402d" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.848552 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c8fb5597c-2d82q"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.848947 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" podUID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerName="dnsmasq-dns" containerID="cri-o://b376d6f31b26fb3b5d8428d35df403e2d7ba33125152c97ad1261520a5e2f273" gracePeriod=10 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.861419 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-69454fc688-t54w7"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.861729 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-69454fc688-t54w7" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-log" containerID="cri-o://6a378fe14ccec4bc8e63806ae734f9293c3f59cfa0f3b3d5fb967e83eb1a42a1" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.861893 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-69454fc688-t54w7" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-api" containerID="cri-o://ac9c0edaa246fce85ea02c792d4ac1c07afd3d940a4142fe3a468aaa6e107e16" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.876095 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880008 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-server" containerID="cri-o://2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880073 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-server" containerID="cri-o://7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880151 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-updater" containerID="cri-o://a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880191 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-auditor" containerID="cri-o://cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880221 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-replicator" containerID="cri-o://51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880248 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-server" containerID="cri-o://941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880277 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-reaper" containerID="cri-o://8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880306 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-auditor" containerID="cri-o://4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880334 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-replicator" containerID="cri-o://eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880443 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-expirer" containerID="cri-o://3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880491 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="swift-recon-cron" containerID="cri-o://a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880522 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="rsync" containerID="cri-o://a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880716 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-auditor" containerID="cri-o://19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.880784 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-updater" containerID="cri-o://4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.882522 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-replicator" containerID="cri-o://caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.903621 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.903938 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-log" containerID="cri-o://04a7a8335ee4dd2cc59886e31042cdc9337381a24ca52a361f5d5f35b6bdead5" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.904351 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-metadata" containerID="cri-o://99eb604fda1c1f8beeca26feebc6e7d241d1c8fcbc5022252d4d91c35a9e66a0" gracePeriod=30 Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.913892 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n7dp\" (UniqueName: \"kubernetes.io/projected/7b3d3641-40ca-4391-b7f2-811a85e9c99e-kube-api-access-4n7dp\") pod \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.916513 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-config\") pod \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.916585 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.916710 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdb-rundir\") pod \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.916763 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-combined-ca-bundle\") pod \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.916803 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-scripts\") pod \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.916975 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-metrics-certs-tls-certs\") pod \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.917068 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdbserver-sb-tls-certs\") pod \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\" (UID: \"7b3d3641-40ca-4391-b7f2-811a85e9c99e\") " Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.921319 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-config" (OuterVolumeSpecName: "config") pod "7b3d3641-40ca-4391-b7f2-811a85e9c99e" (UID: "7b3d3641-40ca-4391-b7f2-811a85e9c99e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.921383 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-scripts" (OuterVolumeSpecName: "scripts") pod "7b3d3641-40ca-4391-b7f2-811a85e9c99e" (UID: "7b3d3641-40ca-4391-b7f2-811a85e9c99e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.922155 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "7b3d3641-40ca-4391-b7f2-811a85e9c99e" (UID: "7b3d3641-40ca-4391-b7f2-811a85e9c99e"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.966004 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "7b3d3641-40ca-4391-b7f2-811a85e9c99e" (UID: "7b3d3641-40ca-4391-b7f2-811a85e9c99e"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:04 crc kubenswrapper[4946]: I1203 07:16:04.983731 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.001499 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.001778 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-log" containerID="cri-o://6c6cec4e34e14e36e892d11434e257757b6502d3fb6f96b9a2e74b1f7a78898a" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.002374 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-api" containerID="cri-o://434d3710c0181e612e1826bfbdad35d34bd77749c41c0c2a77720e391d7d862e" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.027107 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.027168 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.027181 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.027191 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b3d3641-40ca-4391-b7f2-811a85e9c99e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.069982 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.095876 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b3d3641-40ca-4391-b7f2-811a85e9c99e-kube-api-access-4n7dp" (OuterVolumeSpecName: "kube-api-access-4n7dp") pod "7b3d3641-40ca-4391-b7f2-811a85e9c99e" (UID: "7b3d3641-40ca-4391-b7f2-811a85e9c99e"). InnerVolumeSpecName "kube-api-access-4n7dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.121601 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 03 07:16:05 crc kubenswrapper[4946]: E1203 07:16:05.144139 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.144493 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.144524 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n7dp\" (UniqueName: \"kubernetes.io/projected/7b3d3641-40ca-4391-b7f2-811a85e9c99e-kube-api-access-4n7dp\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: E1203 07:16:05.144609 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 07:16:05 crc kubenswrapper[4946]: E1203 07:16:05.144663 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data podName:2f821218-d2a4-4f97-a924-199c056c7c55 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:06.144643118 +0000 UTC m=+1558.941333217 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data") pod "rabbitmq-cell1-server-0" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55") : configmap "rabbitmq-cell1-config-data" not found Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.144692 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-gtzjg"] Dec 03 07:16:05 crc kubenswrapper[4946]: E1203 07:16:05.145784 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data podName:854575dc-d77d-43fe-a7ff-495ef76ab456 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:07.145725157 +0000 UTC m=+1559.942415266 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data") pod "rabbitmq-server-0" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456") : configmap "rabbitmq-config-data" not found Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.149477 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" podUID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.196:5353: connect: connection refused" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.192383 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b3d3641-40ca-4391-b7f2-811a85e9c99e" (UID: "7b3d3641-40ca-4391-b7f2-811a85e9c99e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.208155 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8260f89c-239d-42e6-b187-c912013d5eca/ovsdbserver-nb/0.log" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.208248 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.211346 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-gtzjg"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.256094 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.266621 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6786b69d4-clggf"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.267175 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6786b69d4-clggf" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api-log" containerID="cri-o://29a3b6fe73a28a60ca8a3174c922f9babad002d47efe3d41eecf7fee894fc739" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.267363 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6786b69d4-clggf" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api" containerID="cri-o://364174d07480a501e011d47109a2fdb307a389eee25a760b44b2337057f852ec" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.273596 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f06f-account-create-update-htsj6"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.291534 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-559fcd84cc-lhklt"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.291924 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-559fcd84cc-lhklt" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerName="barbican-worker-log" containerID="cri-o://27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.293430 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-559fcd84cc-lhklt" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerName="barbican-worker" containerID="cri-o://ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.304138 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f06f-account-create-update-htsj6"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.337601 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="854575dc-d77d-43fe-a7ff-495ef76ab456" containerName="rabbitmq" containerID="cri-o://2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32" gracePeriod=604800 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.337653 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7f55c455cd-jdhgw"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.338276 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerName="barbican-keystone-listener-log" containerID="cri-o://444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.338341 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerName="barbican-keystone-listener" containerID="cri-o://4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373015 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" probeResult="failure" output=< Dec 03 07:16:05 crc kubenswrapper[4946]: cat: /var/run/openvswitch/ovs-vswitchd.pid: No such file or directory Dec 03 07:16:05 crc kubenswrapper[4946]: ERROR - Failed to get pid for ovs-vswitchd, exit status: 0 Dec 03 07:16:05 crc kubenswrapper[4946]: > Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373319 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-config\") pod \"8260f89c-239d-42e6-b187-c912013d5eca\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373354 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-scripts\") pod \"8260f89c-239d-42e6-b187-c912013d5eca\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373425 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvjls\" (UniqueName: \"kubernetes.io/projected/8260f89c-239d-42e6-b187-c912013d5eca-kube-api-access-vvjls\") pod \"8260f89c-239d-42e6-b187-c912013d5eca\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373560 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-combined-ca-bundle\") pod \"8260f89c-239d-42e6-b187-c912013d5eca\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373616 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-metrics-certs-tls-certs\") pod \"8260f89c-239d-42e6-b187-c912013d5eca\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373652 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8260f89c-239d-42e6-b187-c912013d5eca-ovsdb-rundir\") pod \"8260f89c-239d-42e6-b187-c912013d5eca\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373689 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-ovsdbserver-nb-tls-certs\") pod \"8260f89c-239d-42e6-b187-c912013d5eca\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.373775 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"8260f89c-239d-42e6-b187-c912013d5eca\" (UID: \"8260f89c-239d-42e6-b187-c912013d5eca\") " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.377379 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-scripts" (OuterVolumeSpecName: "scripts") pod "8260f89c-239d-42e6-b187-c912013d5eca" (UID: "8260f89c-239d-42e6-b187-c912013d5eca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.390255 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8260f89c-239d-42e6-b187-c912013d5eca-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "8260f89c-239d-42e6-b187-c912013d5eca" (UID: "8260f89c-239d-42e6-b187-c912013d5eca"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.390380 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" containerID="cri-o://019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.391196 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.398269 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-config" (OuterVolumeSpecName: "config") pod "8260f89c-239d-42e6-b187-c912013d5eca" (UID: "8260f89c-239d-42e6-b187-c912013d5eca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.398847 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="04384280-d303-475d-8abc-c9c957db7fd9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://c1b84099e3ff7baff4b0d71a61709731ec4d0a9f91be8cac1668b867cdc1f2bf" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.429276 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8260f89c-239d-42e6-b187-c912013d5eca-kube-api-access-vvjls" (OuterVolumeSpecName: "kube-api-access-vvjls") pod "8260f89c-239d-42e6-b187-c912013d5eca" (UID: "8260f89c-239d-42e6-b187-c912013d5eca"). InnerVolumeSpecName "kube-api-access-vvjls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.429266 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" containerName="galera" containerID="cri-o://f019b96d17fb751f8b25510337f077e065922ffd4d5f69477d62724118532891" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.431104 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.444683 4946 generic.go:334] "Generic (PLEG): container finished" podID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerID="86788e20897ae74696026dd8776710b0bda123a3f5dc62f4eebb35926ddd9811" exitCode=143 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.444816 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bd1d9c96-a0f4-43db-a686-2b320c1948ae","Type":"ContainerDied","Data":"86788e20897ae74696026dd8776710b0bda123a3f5dc62f4eebb35926ddd9811"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.446994 4946 generic.go:334] "Generic (PLEG): container finished" podID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerID="ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.447070 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60cdd367-2981-42ad-a896-23ceee7ac34e","Type":"ContainerDied","Data":"ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.449776 4946 generic.go:334] "Generic (PLEG): container finished" podID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerID="33b47ed2f51fb1905563dad28b7eecdde497b32b4bd342ccb3247ecc1b377e2f" exitCode=143 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.449833 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7758850c-e98d-4bb6-bd84-31584a9bcdb6","Type":"ContainerDied","Data":"33b47ed2f51fb1905563dad28b7eecdde497b32b4bd342ccb3247ecc1b377e2f"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.457112 4946 generic.go:334] "Generic (PLEG): container finished" podID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerID="6c6cec4e34e14e36e892d11434e257757b6502d3fb6f96b9a2e74b1f7a78898a" exitCode=143 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.457166 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df","Type":"ContainerDied","Data":"6c6cec4e34e14e36e892d11434e257757b6502d3fb6f96b9a2e74b1f7a78898a"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.458481 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "8260f89c-239d-42e6-b187-c912013d5eca" (UID: "8260f89c-239d-42e6-b187-c912013d5eca"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.459578 4946 generic.go:334] "Generic (PLEG): container finished" podID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerID="6a378fe14ccec4bc8e63806ae734f9293c3f59cfa0f3b3d5fb967e83eb1a42a1" exitCode=143 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.459651 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69454fc688-t54w7" event={"ID":"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541","Type":"ContainerDied","Data":"6a378fe14ccec4bc8e63806ae734f9293c3f59cfa0f3b3d5fb967e83eb1a42a1"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.461855 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8260f89c-239d-42e6-b187-c912013d5eca/ovsdbserver-nb/0.log" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.461918 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8260f89c-239d-42e6-b187-c912013d5eca","Type":"ContainerDied","Data":"cd6cd0524b7074097a1a5912f376575963143b3d36d4647375c6922450a76242"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.461944 4946 scope.go:117] "RemoveContainer" containerID="bbdd199402f5e0e7ad481bf616fe43d3edcdb9ecdbe811804101a2daa767e736" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.462060 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 07:16:05 crc kubenswrapper[4946]: E1203 07:16:05.464976 4946 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 03 07:16:05 crc kubenswrapper[4946]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 03 07:16:05 crc kubenswrapper[4946]: + source /usr/local/bin/container-scripts/functions Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNBridge=br-int Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNRemote=tcp:localhost:6642 Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNEncapType=geneve Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNAvailabilityZones= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ EnableChassisAsGateway=true Dec 03 07:16:05 crc kubenswrapper[4946]: ++ PhysicalNetworks= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNHostName= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 03 07:16:05 crc kubenswrapper[4946]: ++ ovs_dir=/var/lib/openvswitch Dec 03 07:16:05 crc kubenswrapper[4946]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 03 07:16:05 crc kubenswrapper[4946]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 03 07:16:05 crc kubenswrapper[4946]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 03 07:16:05 crc kubenswrapper[4946]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 03 07:16:05 crc kubenswrapper[4946]: + sleep 0.5 Dec 03 07:16:05 crc kubenswrapper[4946]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 03 07:16:05 crc kubenswrapper[4946]: + cleanup_ovsdb_server_semaphore Dec 03 07:16:05 crc kubenswrapper[4946]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 03 07:16:05 crc kubenswrapper[4946]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 03 07:16:05 crc kubenswrapper[4946]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-sk56f" message=< Dec 03 07:16:05 crc kubenswrapper[4946]: Exiting ovsdb-server (5) [ OK ] Dec 03 07:16:05 crc kubenswrapper[4946]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 03 07:16:05 crc kubenswrapper[4946]: + source /usr/local/bin/container-scripts/functions Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNBridge=br-int Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNRemote=tcp:localhost:6642 Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNEncapType=geneve Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNAvailabilityZones= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ EnableChassisAsGateway=true Dec 03 07:16:05 crc kubenswrapper[4946]: ++ PhysicalNetworks= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNHostName= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 03 07:16:05 crc kubenswrapper[4946]: ++ ovs_dir=/var/lib/openvswitch Dec 03 07:16:05 crc kubenswrapper[4946]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 03 07:16:05 crc kubenswrapper[4946]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 03 07:16:05 crc kubenswrapper[4946]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 03 07:16:05 crc kubenswrapper[4946]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 03 07:16:05 crc kubenswrapper[4946]: + sleep 0.5 Dec 03 07:16:05 crc kubenswrapper[4946]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 03 07:16:05 crc kubenswrapper[4946]: + cleanup_ovsdb_server_semaphore Dec 03 07:16:05 crc kubenswrapper[4946]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 03 07:16:05 crc kubenswrapper[4946]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 03 07:16:05 crc kubenswrapper[4946]: > Dec 03 07:16:05 crc kubenswrapper[4946]: E1203 07:16:05.465015 4946 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 03 07:16:05 crc kubenswrapper[4946]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 03 07:16:05 crc kubenswrapper[4946]: + source /usr/local/bin/container-scripts/functions Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNBridge=br-int Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNRemote=tcp:localhost:6642 Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNEncapType=geneve Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNAvailabilityZones= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ EnableChassisAsGateway=true Dec 03 07:16:05 crc kubenswrapper[4946]: ++ PhysicalNetworks= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ OVNHostName= Dec 03 07:16:05 crc kubenswrapper[4946]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 03 07:16:05 crc kubenswrapper[4946]: ++ ovs_dir=/var/lib/openvswitch Dec 03 07:16:05 crc kubenswrapper[4946]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 03 07:16:05 crc kubenswrapper[4946]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 03 07:16:05 crc kubenswrapper[4946]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 03 07:16:05 crc kubenswrapper[4946]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 03 07:16:05 crc kubenswrapper[4946]: + sleep 0.5 Dec 03 07:16:05 crc kubenswrapper[4946]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 03 07:16:05 crc kubenswrapper[4946]: + cleanup_ovsdb_server_semaphore Dec 03 07:16:05 crc kubenswrapper[4946]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 03 07:16:05 crc kubenswrapper[4946]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 03 07:16:05 crc kubenswrapper[4946]: > pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" containerID="cri-o://2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.465049 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" containerID="cri-o://2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.466450 4946 generic.go:334] "Generic (PLEG): container finished" podID="cc385f93-2111-42cf-93be-d4cc82225570" containerID="01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.466495 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86855dfc4c-jdcjc" event={"ID":"cc385f93-2111-42cf-93be-d4cc82225570","Type":"ContainerDied","Data":"01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.468187 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.468398 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e5415d2e-cf73-4dd8-900d-211b957e829f" containerName="nova-scheduler-scheduler" containerID="cri-o://5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.470044 4946 generic.go:334] "Generic (PLEG): container finished" podID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerID="b376d6f31b26fb3b5d8428d35df403e2d7ba33125152c97ad1261520a5e2f273" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.470095 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" event={"ID":"8126945c-b3f3-4fc8-a5eb-553c41a439d5","Type":"ContainerDied","Data":"b376d6f31b26fb3b5d8428d35df403e2d7ba33125152c97ad1261520a5e2f273"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.476909 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7b3d3641-40ca-4391-b7f2-811a85e9c99e/ovsdbserver-sb/0.log" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.476977 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7b3d3641-40ca-4391-b7f2-811a85e9c99e","Type":"ContainerDied","Data":"707e8b7baeee32b7b585df2c1a766119a5228d390ce4b3b00eab5cc7ffd10a48"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.477051 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.477852 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8260f89c-239d-42e6-b187-c912013d5eca-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.477897 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.477911 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.477924 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8260f89c-239d-42e6-b187-c912013d5eca-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.477934 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvjls\" (UniqueName: \"kubernetes.io/projected/8260f89c-239d-42e6-b187-c912013d5eca-kube-api-access-vvjls\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.514768 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.514874 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.514944 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515187 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515240 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515294 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515351 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515407 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515470 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515522 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515576 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515627 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515709 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515918 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.515994 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.516215 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.516982 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.517067 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.517148 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.517340 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.517616 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.517997 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.518097 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.518487 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.531610 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fh88p_df3b0afc-1419-44ce-89f7-b1aca189e145/openstack-network-exporter/0.log" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.531652 4946 generic.go:334] "Generic (PLEG): container finished" podID="df3b0afc-1419-44ce-89f7-b1aca189e145" containerID="2eb06361aba92c8a4af213131c660c362c2c6949de5340c6fe246ad8413ed5fd" exitCode=2 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.531837 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fh88p" event={"ID":"df3b0afc-1419-44ce-89f7-b1aca189e145","Type":"ContainerDied","Data":"2eb06361aba92c8a4af213131c660c362c2c6949de5340c6fe246ad8413ed5fd"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.531890 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8260f89c-239d-42e6-b187-c912013d5eca" (UID: "8260f89c-239d-42e6-b187-c912013d5eca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.535224 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7b3d3641-40ca-4391-b7f2-811a85e9c99e" (UID: "7b3d3641-40ca-4391-b7f2-811a85e9c99e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.549294 4946 generic.go:334] "Generic (PLEG): container finished" podID="788c3bf1-3291-4324-8a4c-dc06f1027f5b" containerID="4aaa4e05b46fa6c68df1bd9487eadf3885468fbce3baff034433d3bc2c281e0b" exitCode=137 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.553113 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.553919 4946 generic.go:334] "Generic (PLEG): container finished" podID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerID="04a7a8335ee4dd2cc59886e31042cdc9337381a24ca52a361f5d5f35b6bdead5" exitCode=143 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.554908 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"192c2964-d6c9-404e-8cca-cac01c750c5b","Type":"ContainerDied","Data":"04a7a8335ee4dd2cc59886e31042cdc9337381a24ca52a361f5d5f35b6bdead5"} Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.555155 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xcjbh" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="registry-server" containerID="cri-o://d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e" gracePeriod=2 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.582160 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "7b3d3641-40ca-4391-b7f2-811a85e9c99e" (UID: "7b3d3641-40ca-4391-b7f2-811a85e9c99e"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.587816 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancec521-account-delete-d62ms"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.601805 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.601840 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.601850 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b3d3641-40ca-4391-b7f2-811a85e9c99e-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.601861 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.644802 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0190a925-4703-4ae4-8ab4-150aef815c52" path="/var/lib/kubelet/pods/0190a925-4703-4ae4-8ab4-150aef815c52/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.645023 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="2f821218-d2a4-4f97-a924-199c056c7c55" containerName="rabbitmq" containerID="cri-o://7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e" gracePeriod=604800 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.645432 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a" path="/var/lib/kubelet/pods/5c0c0a3b-3342-4ed5-a1c2-d5ecf7c7494a/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.646636 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cba79e5-c505-4c72-9609-48838c0f59c5" path="/var/lib/kubelet/pods/6cba79e5-c505-4c72-9609-48838c0f59c5/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.647775 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="778a6a04-4fc8-46fc-8f83-9d3625eb93d8" path="/var/lib/kubelet/pods/778a6a04-4fc8-46fc-8f83-9d3625eb93d8/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.650751 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e368226-3cfe-42a1-b2a3-52a53d227807" path="/var/lib/kubelet/pods/8e368226-3cfe-42a1-b2a3-52a53d227807/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.651554 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a59904e6-2de4-4bc6-b0b5-6f578470b7d6" path="/var/lib/kubelet/pods/a59904e6-2de4-4bc6-b0b5-6f578470b7d6/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.653516 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8760772-02a5-4231-9687-b0cd0f34f5c2" path="/var/lib/kubelet/pods/a8760772-02a5-4231-9687-b0cd0f34f5c2/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.654056 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24f9dfa-7c1a-4361-88d8-52b9a0ea559e" path="/var/lib/kubelet/pods/b24f9dfa-7c1a-4361-88d8-52b9a0ea559e/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.658998 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be8091e1-4683-4e98-87bc-ecd638b6cbfb" path="/var/lib/kubelet/pods/be8091e1-4683-4e98-87bc-ecd638b6cbfb/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.659505 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4ad7e1-ae04-4f3a-8594-2897d4052e32" path="/var/lib/kubelet/pods/db4ad7e1-ae04-4f3a-8594-2897d4052e32/volumes" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.716876 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "8260f89c-239d-42e6-b187-c912013d5eca" (UID: "8260f89c-239d-42e6-b187-c912013d5eca"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.810567 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.852037 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xp9sr"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.852643 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xp9sr"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.852668 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.852691 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.852705 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8vff8"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.852721 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8vff8"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.852930 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" containerName="nova-cell1-conductor-conductor" containerID="cri-o://5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.853113 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="0e0b4114-0121-48cc-9439-bc5d350aeb44" containerName="nova-cell0-conductor-conductor" containerID="cri-o://d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217" gracePeriod=30 Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.883248 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "8260f89c-239d-42e6-b187-c912013d5eca" (UID: "8260f89c-239d-42e6-b187-c912013d5eca"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.889544 4946 scope.go:117] "RemoveContainer" containerID="5b78ba643d512f9707208df0f4f8fcbbe76854d669c9d51da5ef8396964bb755" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.919672 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8260f89c-239d-42e6-b187-c912013d5eca-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.926065 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fh88p_df3b0afc-1419-44ce-89f7-b1aca189e145/openstack-network-exporter/0.log" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.926135 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.953283 4946 scope.go:117] "RemoveContainer" containerID="dbf1d62862ef8e193ae4fb4237e071cfc9defb2902e956f02581df96ce0e9858" Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.977879 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron6e16-account-delete-snvqr"] Dec 03 07:16:05 crc kubenswrapper[4946]: I1203 07:16:05.979447 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.004825 4946 scope.go:117] "RemoveContainer" containerID="306f0aa54817d7e3ac475556197a16b148ab3536c5c0e1a6f77075f7234d7b98" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.004976 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.023603 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024152 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-sb\") pod \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024209 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2skq\" (UniqueName: \"kubernetes.io/projected/8126945c-b3f3-4fc8-a5eb-553c41a439d5-kube-api-access-j2skq\") pod \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024266 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-svc\") pod \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024328 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-swift-storage-0\") pod \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024362 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-combined-ca-bundle\") pod \"df3b0afc-1419-44ce-89f7-b1aca189e145\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024384 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-metrics-certs-tls-certs\") pod \"df3b0afc-1419-44ce-89f7-b1aca189e145\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024447 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpdsw\" (UniqueName: \"kubernetes.io/projected/df3b0afc-1419-44ce-89f7-b1aca189e145-kube-api-access-jpdsw\") pod \"df3b0afc-1419-44ce-89f7-b1aca189e145\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024472 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df3b0afc-1419-44ce-89f7-b1aca189e145-config\") pod \"df3b0afc-1419-44ce-89f7-b1aca189e145\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024508 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-config\") pod \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024526 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovn-rundir\") pod \"df3b0afc-1419-44ce-89f7-b1aca189e145\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024546 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovs-rundir\") pod \"df3b0afc-1419-44ce-89f7-b1aca189e145\" (UID: \"df3b0afc-1419-44ce-89f7-b1aca189e145\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.024561 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-nb\") pod \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\" (UID: \"8126945c-b3f3-4fc8-a5eb-553c41a439d5\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.033587 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df3b0afc-1419-44ce-89f7-b1aca189e145-config" (OuterVolumeSpecName: "config") pod "df3b0afc-1419-44ce-89f7-b1aca189e145" (UID: "df3b0afc-1419-44ce-89f7-b1aca189e145"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.033914 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "df3b0afc-1419-44ce-89f7-b1aca189e145" (UID: "df3b0afc-1419-44ce-89f7-b1aca189e145"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.033968 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "df3b0afc-1419-44ce-89f7-b1aca189e145" (UID: "df3b0afc-1419-44ce-89f7-b1aca189e145"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.039113 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder5125-account-delete-lrpcc"] Dec 03 07:16:06 crc kubenswrapper[4946]: W1203 07:16:06.042630 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6e24ccc_b386_4d6e_9c67_cc95e47cafab.slice/crio-d6ca54dc63480a9b69f63634dde246e53accc0de903e525d1b3d24fce74ad4dd WatchSource:0}: Error finding container d6ca54dc63480a9b69f63634dde246e53accc0de903e525d1b3d24fce74ad4dd: Status 404 returned error can't find the container with id d6ca54dc63480a9b69f63634dde246e53accc0de903e525d1b3d24fce74ad4dd Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.048922 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement0960-account-delete-pc5x5"] Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.056441 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df3b0afc-1419-44ce-89f7-b1aca189e145-kube-api-access-jpdsw" (OuterVolumeSpecName: "kube-api-access-jpdsw") pod "df3b0afc-1419-44ce-89f7-b1aca189e145" (UID: "df3b0afc-1419-44ce-89f7-b1aca189e145"). InnerVolumeSpecName "kube-api-access-jpdsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.057022 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8126945c-b3f3-4fc8-a5eb-553c41a439d5-kube-api-access-j2skq" (OuterVolumeSpecName: "kube-api-access-j2skq") pod "8126945c-b3f3-4fc8-a5eb-553c41a439d5" (UID: "8126945c-b3f3-4fc8-a5eb-553c41a439d5"). InnerVolumeSpecName "kube-api-access-j2skq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.065778 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicaneae6-account-delete-l9467"] Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.114698 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.123464 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.128823 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpdsw\" (UniqueName: \"kubernetes.io/projected/df3b0afc-1419-44ce-89f7-b1aca189e145-kube-api-access-jpdsw\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.128845 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df3b0afc-1419-44ce-89f7-b1aca189e145-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.128854 4946 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.128864 4946 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/df3b0afc-1419-44ce-89f7-b1aca189e145-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.128872 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2skq\" (UniqueName: \"kubernetes.io/projected/8126945c-b3f3-4fc8-a5eb-553c41a439d5-kube-api-access-j2skq\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.168148 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.229486 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-combined-ca-bundle\") pod \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.229592 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcb4v\" (UniqueName: \"kubernetes.io/projected/788c3bf1-3291-4324-8a4c-dc06f1027f5b-kube-api-access-hcb4v\") pod \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.229726 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config\") pod \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.229817 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config-secret\") pod \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\" (UID: \"788c3bf1-3291-4324-8a4c-dc06f1027f5b\") " Dec 03 07:16:06 crc kubenswrapper[4946]: E1203 07:16:06.230189 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 07:16:06 crc kubenswrapper[4946]: E1203 07:16:06.230231 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data podName:2f821218-d2a4-4f97-a924-199c056c7c55 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:08.230219221 +0000 UTC m=+1561.026909330 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data") pod "rabbitmq-cell1-server-0" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55") : configmap "rabbitmq-cell1-config-data" not found Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.234364 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788c3bf1-3291-4324-8a4c-dc06f1027f5b-kube-api-access-hcb4v" (OuterVolumeSpecName: "kube-api-access-hcb4v") pod "788c3bf1-3291-4324-8a4c-dc06f1027f5b" (UID: "788c3bf1-3291-4324-8a4c-dc06f1027f5b"). InnerVolumeSpecName "kube-api-access-hcb4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.243295 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.291829 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8126945c-b3f3-4fc8-a5eb-553c41a439d5" (UID: "8126945c-b3f3-4fc8-a5eb-553c41a439d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.341320 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df3b0afc-1419-44ce-89f7-b1aca189e145" (UID: "df3b0afc-1419-44ce-89f7-b1aca189e145"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.343885 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.343946 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.343965 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcb4v\" (UniqueName: \"kubernetes.io/projected/788c3bf1-3291-4324-8a4c-dc06f1027f5b-kube-api-access-hcb4v\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.349207 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-8587b466c5-9tcks"] Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.350645 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-8587b466c5-9tcks" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerName="proxy-httpd" containerID="cri-o://2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1" gracePeriod=30 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.350973 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-8587b466c5-9tcks" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerName="proxy-server" containerID="cri-o://73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8" gracePeriod=30 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.419710 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "788c3bf1-3291-4324-8a4c-dc06f1027f5b" (UID: "788c3bf1-3291-4324-8a4c-dc06f1027f5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.445467 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8czkh\" (UniqueName: \"kubernetes.io/projected/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-kube-api-access-8czkh\") pod \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.445565 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-utilities\") pod \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.445724 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-catalog-content\") pod \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\" (UID: \"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb\") " Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.446605 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.448451 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-utilities" (OuterVolumeSpecName: "utilities") pod "6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" (UID: "6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.461264 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8126945c-b3f3-4fc8-a5eb-553c41a439d5" (UID: "8126945c-b3f3-4fc8-a5eb-553c41a439d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.489006 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-kube-api-access-8czkh" (OuterVolumeSpecName: "kube-api-access-8czkh") pod "6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" (UID: "6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb"). InnerVolumeSpecName "kube-api-access-8czkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.545629 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "788c3bf1-3291-4324-8a4c-dc06f1027f5b" (UID: "788c3bf1-3291-4324-8a4c-dc06f1027f5b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.552873 4946 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.552912 4946 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.552924 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8czkh\" (UniqueName: \"kubernetes.io/projected/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-kube-api-access-8czkh\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.552934 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.572097 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8126945c-b3f3-4fc8-a5eb-553c41a439d5" (UID: "8126945c-b3f3-4fc8-a5eb-553c41a439d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.620388 4946 generic.go:334] "Generic (PLEG): container finished" podID="761f2199-bf20-487a-a18d-172c65a4b515" containerID="29a3b6fe73a28a60ca8a3174c922f9babad002d47efe3d41eecf7fee894fc739" exitCode=143 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.620462 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6786b69d4-clggf" event={"ID":"761f2199-bf20-487a-a18d-172c65a4b515","Type":"ContainerDied","Data":"29a3b6fe73a28a60ca8a3174c922f9babad002d47efe3d41eecf7fee894fc739"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.641439 4946 generic.go:334] "Generic (PLEG): container finished" podID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerID="d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e" exitCode=0 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.641526 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjbh" event={"ID":"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb","Type":"ContainerDied","Data":"d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.641541 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcjbh" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.641563 4946 scope.go:117] "RemoveContainer" containerID="d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.641553 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjbh" event={"ID":"6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb","Type":"ContainerDied","Data":"69e66f0ed3549111e29aa3b76ae5254f6c3de0f690272416b58064d884a00d41"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.644791 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicaneae6-account-delete-l9467" event={"ID":"85a6dcc4-23cc-4f57-8714-06d0ef25c680","Type":"ContainerStarted","Data":"1e2834fdf3e1e0b071ee3d2989261ba331c5000d483d4e7b2af82af3bbc9b178"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.648009 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" (UID: "6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.648881 4946 generic.go:334] "Generic (PLEG): container finished" podID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerID="27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060" exitCode=143 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.648950 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-559fcd84cc-lhklt" event={"ID":"26ac177b-b889-43e7-888e-8ed02fb3ac72","Type":"ContainerDied","Data":"27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.654197 4946 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.654227 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.655495 4946 generic.go:334] "Generic (PLEG): container finished" podID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerID="444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510" exitCode=143 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.655573 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" event={"ID":"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e","Type":"ContainerDied","Data":"444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.657092 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fh88p_df3b0afc-1419-44ce-89f7-b1aca189e145/openstack-network-exporter/0.log" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.657147 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fh88p" event={"ID":"df3b0afc-1419-44ce-89f7-b1aca189e145","Type":"ContainerDied","Data":"2dbfac30a23b90732cb1205a995f0464563b345021e6f412353da20ae2cb4121"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.657228 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fh88p" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.661358 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron6e16-account-delete-snvqr" event={"ID":"b6e24ccc-b386-4d6e-9c67-cc95e47cafab","Type":"ContainerStarted","Data":"d6ca54dc63480a9b69f63634dde246e53accc0de903e525d1b3d24fce74ad4dd"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.662824 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0960-account-delete-pc5x5" event={"ID":"572a50eb-42e9-4897-9925-4073b5f6a35d","Type":"ContainerStarted","Data":"1a891739c6bad23592f03f5fbaa0c50d11373da912e159446c140e5d10c4e675"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.664199 4946 generic.go:334] "Generic (PLEG): container finished" podID="0e57ed09-f7ce-403e-9cb5-676ae1c64117" containerID="cccc8d922dd34c8481eefd183a382905de547dc11a0e346c9e3f2854da1c3218" exitCode=0 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.664244 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancec521-account-delete-d62ms" event={"ID":"0e57ed09-f7ce-403e-9cb5-676ae1c64117","Type":"ContainerDied","Data":"cccc8d922dd34c8481eefd183a382905de547dc11a0e346c9e3f2854da1c3218"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.664259 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancec521-account-delete-d62ms" event={"ID":"0e57ed09-f7ce-403e-9cb5-676ae1c64117","Type":"ContainerStarted","Data":"2e115a720f507036e37b79865616b059c331fb01e03b2871ece60f2c8778b9ee"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.680204 4946 generic.go:334] "Generic (PLEG): container finished" podID="04384280-d303-475d-8abc-c9c957db7fd9" containerID="c1b84099e3ff7baff4b0d71a61709731ec4d0a9f91be8cac1668b867cdc1f2bf" exitCode=0 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.680298 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"04384280-d303-475d-8abc-c9c957db7fd9","Type":"ContainerDied","Data":"c1b84099e3ff7baff4b0d71a61709731ec4d0a9f91be8cac1668b867cdc1f2bf"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.685502 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5125-account-delete-lrpcc" event={"ID":"baf286a1-0b76-4e66-817c-7d7ce84edaa2","Type":"ContainerStarted","Data":"03ef8c2e844685ca4de40a06f36b963e777a38ea495584831ab51b98dd3d0a77"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.694159 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.698980 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8126945c-b3f3-4fc8-a5eb-553c41a439d5" (UID: "8126945c-b3f3-4fc8-a5eb-553c41a439d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.701360 4946 generic.go:334] "Generic (PLEG): container finished" podID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" exitCode=0 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.701521 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sk56f" event={"ID":"f8c1e670-a923-4bf5-a884-23644b2e52aa","Type":"ContainerDied","Data":"2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.708962 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.709772 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8fb5597c-2d82q" event={"ID":"8126945c-b3f3-4fc8-a5eb-553c41a439d5","Type":"ContainerDied","Data":"b9ad8d184d12d742da68ba2d1c8fd01b68c215fe4f2e6d11c071a45476078824"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.723582 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-config" (OuterVolumeSpecName: "config") pod "8126945c-b3f3-4fc8-a5eb-553c41a439d5" (UID: "8126945c-b3f3-4fc8-a5eb-553c41a439d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.737323 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a" exitCode=0 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.737348 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826" exitCode=0 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.737386 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.737412 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.740696 4946 generic.go:334] "Generic (PLEG): container finished" podID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" containerID="f019b96d17fb751f8b25510337f077e065922ffd4d5f69477d62724118532891" exitCode=0 Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.740720 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345","Type":"ContainerDied","Data":"f019b96d17fb751f8b25510337f077e065922ffd4d5f69477d62724118532891"} Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.756294 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "df3b0afc-1419-44ce-89f7-b1aca189e145" (UID: "df3b0afc-1419-44ce-89f7-b1aca189e145"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.762069 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.768766 4946 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8126945c-b3f3-4fc8-a5eb-553c41a439d5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.768779 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df3b0afc-1419-44ce-89f7-b1aca189e145-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.775026 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "788c3bf1-3291-4324-8a4c-dc06f1027f5b" (UID: "788c3bf1-3291-4324-8a4c-dc06f1027f5b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.873244 4946 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/788c3bf1-3291-4324-8a4c-dc06f1027f5b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:06 crc kubenswrapper[4946]: I1203 07:16:06.922368 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.161:8776/healthcheck\": read tcp 10.217.0.2:41690->10.217.0.161:8776: read: connection reset by peer" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.012539 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi2e82-account-delete-77lgb"] Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.025368 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0a5c7-account-delete-j2fhk"] Dec 03 07:16:07 crc kubenswrapper[4946]: W1203 07:16:07.066009 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4c6455_fec3_41e2_bf0a_3595e34d63aa.slice/crio-c5dd963b92e27b3d88ed547b1a50b5d06468216e8e20c7b71c78967617c31a10 WatchSource:0}: Error finding container c5dd963b92e27b3d88ed547b1a50b5d06468216e8e20c7b71c78967617c31a10: Status 404 returned error can't find the container with id c5dd963b92e27b3d88ed547b1a50b5d06468216e8e20c7b71c78967617c31a10 Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.174878 4946 scope.go:117] "RemoveContainer" containerID="3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee" Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.180670 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.180727 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data podName:854575dc-d77d-43fe-a7ff-495ef76ab456 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:11.180712944 +0000 UTC m=+1563.977403043 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data") pod "rabbitmq-server-0" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456") : configmap "rabbitmq-config-data" not found Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.236054 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf is running failed: container process not found" containerID="5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.236275 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf is running failed: container process not found" containerID="5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.236457 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf is running failed: container process not found" containerID="5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.236483 4946 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" containerName="nova-cell1-conductor-conductor" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.260177 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.281910 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-combined-ca-bundle\") pod \"04384280-d303-475d-8abc-c9c957db7fd9\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.281955 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-vencrypt-tls-certs\") pod \"04384280-d303-475d-8abc-c9c957db7fd9\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.282005 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d79p9\" (UniqueName: \"kubernetes.io/projected/04384280-d303-475d-8abc-c9c957db7fd9-kube-api-access-d79p9\") pod \"04384280-d303-475d-8abc-c9c957db7fd9\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.282082 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-config-data\") pod \"04384280-d303-475d-8abc-c9c957db7fd9\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.282149 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-nova-novncproxy-tls-certs\") pod \"04384280-d303-475d-8abc-c9c957db7fd9\" (UID: \"04384280-d303-475d-8abc-c9c957db7fd9\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.291615 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04384280-d303-475d-8abc-c9c957db7fd9-kube-api-access-d79p9" (OuterVolumeSpecName: "kube-api-access-d79p9") pod "04384280-d303-475d-8abc-c9c957db7fd9" (UID: "04384280-d303-475d-8abc-c9c957db7fd9"). InnerVolumeSpecName "kube-api-access-d79p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.329326 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04384280-d303-475d-8abc-c9c957db7fd9" (UID: "04384280-d303-475d-8abc-c9c957db7fd9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.371939 4946 scope.go:117] "RemoveContainer" containerID="6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.384682 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-config-data" (OuterVolumeSpecName: "config-data") pod "04384280-d303-475d-8abc-c9c957db7fd9" (UID: "04384280-d303-475d-8abc-c9c957db7fd9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.387723 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.387762 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d79p9\" (UniqueName: \"kubernetes.io/projected/04384280-d303-475d-8abc-c9c957db7fd9-kube-api-access-d79p9\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.387772 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.394440 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.417080 4946 scope.go:117] "RemoveContainer" containerID="d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e" Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.417174 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.417200 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c8fb5597c-2d82q"] Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.426072 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.426295 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e\": container with ID starting with d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e not found: ID does not exist" containerID="d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.426319 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e"} err="failed to get container status \"d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e\": rpc error: code = NotFound desc = could not find container \"d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e\": container with ID starting with d02b9036f7fd9064ba6578f89137f068d35407ea224773c1919f42c96bced05e not found: ID does not exist" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.426339 4946 scope.go:117] "RemoveContainer" containerID="3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee" Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.429922 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee\": container with ID starting with 3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee not found: ID does not exist" containerID="3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.429968 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee"} err="failed to get container status \"3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee\": rpc error: code = NotFound desc = could not find container \"3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee\": container with ID starting with 3fcae49f491331555ca711b4ce0d68e7af6c6cf7fcf81a54f8552c7082a772ee not found: ID does not exist" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.429995 4946 scope.go:117] "RemoveContainer" containerID="6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7" Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.430843 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7\": container with ID starting with 6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7 not found: ID does not exist" containerID="6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.430880 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7"} err="failed to get container status \"6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7\": rpc error: code = NotFound desc = could not find container \"6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7\": container with ID starting with 6e0adffa41bcf0965f42c38c4869c9de149f3f00924acd8d7676d82c4d9b4dc7 not found: ID does not exist" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.430909 4946 scope.go:117] "RemoveContainer" containerID="2eb06361aba92c8a4af213131c660c362c2c6949de5340c6fe246ad8413ed5fd" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.431253 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.432176 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 03 07:16:07 crc kubenswrapper[4946]: E1203 07:16:07.432225 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="ovn-northd" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.437223 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c8fb5597c-2d82q"] Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.449834 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xcjbh"] Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.454528 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xcjbh"] Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.471518 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-fh88p"] Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.479582 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-fh88p"] Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.489371 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-combined-ca-bundle\") pod \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.490385 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-galera-tls-certs\") pod \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.491919 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-config-data\") pod \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492039 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-log-httpd\") pod \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492178 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-etc-swift\") pod \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492265 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492368 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-combined-ca-bundle\") pod \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492485 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-internal-tls-certs\") pod \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492586 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-run-httpd\") pod \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492663 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kolla-config\") pod \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492755 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs29c\" (UniqueName: \"kubernetes.io/projected/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kube-api-access-cs29c\") pod \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.492907 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-operator-scripts\") pod \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.493078 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-default\") pod \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.493169 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgc6c\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-kube-api-access-lgc6c\") pod \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.493253 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-generated\") pod \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\" (UID: \"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.493326 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-public-tls-certs\") pod \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\" (UID: \"1e19d50d-c0ed-47b7-af00-6e8d2502f599\") " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.495239 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" (UID: "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.495656 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1e19d50d-c0ed-47b7-af00-6e8d2502f599" (UID: "1e19d50d-c0ed-47b7-af00-6e8d2502f599"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.495967 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" (UID: "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.497548 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" (UID: "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.497989 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" (UID: "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.499118 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1e19d50d-c0ed-47b7-af00-6e8d2502f599" (UID: "1e19d50d-c0ed-47b7-af00-6e8d2502f599"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.501214 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.501243 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.501255 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.501266 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.501275 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e19d50d-c0ed-47b7-af00-6e8d2502f599-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.501283 4946 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.511171 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kube-api-access-cs29c" (OuterVolumeSpecName: "kube-api-access-cs29c") pod "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" (UID: "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345"). InnerVolumeSpecName "kube-api-access-cs29c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.519946 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-kube-api-access-lgc6c" (OuterVolumeSpecName: "kube-api-access-lgc6c") pod "1e19d50d-c0ed-47b7-af00-6e8d2502f599" (UID: "1e19d50d-c0ed-47b7-af00-6e8d2502f599"). InnerVolumeSpecName "kube-api-access-lgc6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.531049 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1e19d50d-c0ed-47b7-af00-6e8d2502f599" (UID: "1e19d50d-c0ed-47b7-af00-6e8d2502f599"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.565508 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" (UID: "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.604757 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs29c\" (UniqueName: \"kubernetes.io/projected/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-kube-api-access-cs29c\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.604790 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgc6c\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-kube-api-access-lgc6c\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.604798 4946 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e19d50d-c0ed-47b7-af00-6e8d2502f599-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.604839 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.621451 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" path="/var/lib/kubelet/pods/6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb/volumes" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.622649 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f4cb0d-e6fd-45dc-be44-40233bad9532" path="/var/lib/kubelet/pods/70f4cb0d-e6fd-45dc-be44-40233bad9532/volumes" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.623301 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788c3bf1-3291-4324-8a4c-dc06f1027f5b" path="/var/lib/kubelet/pods/788c3bf1-3291-4324-8a4c-dc06f1027f5b/volumes" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.624572 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" path="/var/lib/kubelet/pods/7b3d3641-40ca-4391-b7f2-811a85e9c99e/volumes" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.625424 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" path="/var/lib/kubelet/pods/8126945c-b3f3-4fc8-a5eb-553c41a439d5/volumes" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.626712 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8260f89c-239d-42e6-b187-c912013d5eca" path="/var/lib/kubelet/pods/8260f89c-239d-42e6-b187-c912013d5eca/volumes" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.627286 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6109109-dfb1-4b8c-86c7-94d7fe00fed1" path="/var/lib/kubelet/pods/c6109109-dfb1-4b8c-86c7-94d7fe00fed1/volumes" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.627942 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df3b0afc-1419-44ce-89f7-b1aca189e145" path="/var/lib/kubelet/pods/df3b0afc-1419-44ce-89f7-b1aca189e145/volumes" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.730023 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "04384280-d303-475d-8abc-c9c957db7fd9" (UID: "04384280-d303-475d-8abc-c9c957db7fd9"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.790694 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.803524 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.815591 4946 generic.go:334] "Generic (PLEG): container finished" podID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerID="6d8145ae8f7303d9ebe9697e32c04cdfb3a46dd8ce1dc51d2f81b3c03b54e2c8" exitCode=0 Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.820821 4946 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.833248 4946 generic.go:334] "Generic (PLEG): container finished" podID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerID="73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8" exitCode=0 Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.835674 4946 generic.go:334] "Generic (PLEG): container finished" podID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerID="2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1" exitCode=0 Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.835867 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8587b466c5-9tcks" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.863958 4946 generic.go:334] "Generic (PLEG): container finished" podID="baf286a1-0b76-4e66-817c-7d7ce84edaa2" containerID="5f4981a54547ee9f0ce3adc448f87d4c7e86f80240ae2ecacd9efac2c2bb4710" exitCode=0 Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.875454 4946 generic.go:334] "Generic (PLEG): container finished" podID="10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" containerID="5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf" exitCode=0 Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.879826 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapi2e82-account-delete-77lgb" podStartSLOduration=4.879804202 podStartE2EDuration="4.879804202s" podCreationTimestamp="2025-12-03 07:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:16:07.865165199 +0000 UTC m=+1560.661855308" watchObservedRunningTime="2025-12-03 07:16:07.879804202 +0000 UTC m=+1560.676494311" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.946500 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement0960-account-delete-pc5x5" podStartSLOduration=4.946477574 podStartE2EDuration="4.946477574s" podCreationTimestamp="2025-12-03 07:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:16:07.924311268 +0000 UTC m=+1560.721001377" watchObservedRunningTime="2025-12-03 07:16:07.946477574 +0000 UTC m=+1560.743167683" Dec 03 07:16:07 crc kubenswrapper[4946]: I1203 07:16:07.955434 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "04384280-d303-475d-8abc-c9c957db7fd9" (UID: "04384280-d303-475d-8abc-c9c957db7fd9"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:08 crc kubenswrapper[4946]: I1203 07:16:08.025820 4946 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/04384280-d303-475d-8abc-c9c957db7fd9-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:08 crc kubenswrapper[4946]: I1203 07:16:08.044787 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell0a5c7-account-delete-j2fhk" podStartSLOduration=5.044763805 podStartE2EDuration="5.044763805s" podCreationTimestamp="2025-12-03 07:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:16:07.997124095 +0000 UTC m=+1560.793814204" watchObservedRunningTime="2025-12-03 07:16:08.044763805 +0000 UTC m=+1560.841453914" Dec 03 07:16:08 crc kubenswrapper[4946]: E1203 07:16:08.231446 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 03 07:16:08 crc kubenswrapper[4946]: E1203 07:16:08.231749 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data podName:2f821218-d2a4-4f97-a924-199c056c7c55 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:12.231722679 +0000 UTC m=+1565.028412788 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data") pod "rabbitmq-cell1-server-0" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55") : configmap "rabbitmq-cell1-config-data" not found Dec 03 07:16:08 crc kubenswrapper[4946]: I1203 07:16:08.260223 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1e19d50d-c0ed-47b7-af00-6e8d2502f599" (UID: "1e19d50d-c0ed-47b7-af00-6e8d2502f599"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:08 crc kubenswrapper[4946]: I1203 07:16:08.262479 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" (UID: "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.311961 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-69454fc688-t54w7" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-api" probeResult="failure" output="Get \"https://10.217.0.150:8778/\": read tcp 10.217.0.2:44628->10.217.0.150:8778: read: connection reset by peer" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.312256 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-69454fc688-t54w7" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.150:8778/\": read tcp 10.217.0.2:44638->10.217.0.150:8778: read: connection reset by peer" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.339228 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.339252 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.341191 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.344575 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" (UID: "ea1b7f5d-0115-4bd2-8d9c-502eff1ae345"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.440170 4946 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.440198 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.443915 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-config-data" (OuterVolumeSpecName: "config-data") pod "1e19d50d-c0ed-47b7-af00-6e8d2502f599" (UID: "1e19d50d-c0ed-47b7-af00-6e8d2502f599"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.544637 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.546480 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1e19d50d-c0ed-47b7-af00-6e8d2502f599" (UID: "1e19d50d-c0ed-47b7-af00-6e8d2502f599"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.558900 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.560845 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.564994 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.565075 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e5415d2e-cf73-4dd8-900d-211b957e829f" containerName="nova-scheduler-scheduler" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.577201 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": read tcp 10.217.0.2:59654->10.217.0.203:8775: read: connection reset by peer" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.577701 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": read tcp 10.217.0.2:59640->10.217.0.203:8775: read: connection reset by peer" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.583861 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e19d50d-c0ed-47b7-af00-6e8d2502f599" (UID: "1e19d50d-c0ed-47b7-af00-6e8d2502f599"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.591878 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6786b69d4-clggf" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:37044->10.217.0.156:9311: read: connection reset by peer" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.591905 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6786b69d4-clggf" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:37054->10.217.0.156:9311: read: connection reset by peer" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.648208 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.648560 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e19d50d-c0ed-47b7-af00-6e8d2502f599-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.947529 4946 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.355s" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947570 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ea1b7f5d-0115-4bd2-8d9c-502eff1ae345","Type":"ContainerDied","Data":"42caf5eaffec889c4853100be6c5a46874c3730f08900043c8dfa7757d66c6df"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947608 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"04384280-d303-475d-8abc-c9c957db7fd9","Type":"ContainerDied","Data":"e17dd66e133e84e0005889ea0db01e05d4c1c7e86dec7c296cea1a9ac8c3a058"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947622 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi2e82-account-delete-77lgb" event={"ID":"2f4c6455-fec3-41e2-bf0a-3595e34d63aa","Type":"ContainerStarted","Data":"11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947695 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi2e82-account-delete-77lgb" event={"ID":"2f4c6455-fec3-41e2-bf0a-3595e34d63aa","Type":"ContainerStarted","Data":"c5dd963b92e27b3d88ed547b1a50b5d06468216e8e20c7b71c78967617c31a10"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947716 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a3521a3-0f30-4afe-b5f8-efeb6f838e72","Type":"ContainerDied","Data":"6d8145ae8f7303d9ebe9697e32c04cdfb3a46dd8ce1dc51d2f81b3c03b54e2c8"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947727 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a3521a3-0f30-4afe-b5f8-efeb6f838e72","Type":"ContainerDied","Data":"3484bdb32a49685d70086dbee64b43f466364091478c44939ca438f3cbe8d541"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947754 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3484bdb32a49685d70086dbee64b43f466364091478c44939ca438f3cbe8d541" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947765 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8587b466c5-9tcks" event={"ID":"1e19d50d-c0ed-47b7-af00-6e8d2502f599","Type":"ContainerDied","Data":"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947780 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8587b466c5-9tcks" event={"ID":"1e19d50d-c0ed-47b7-af00-6e8d2502f599","Type":"ContainerDied","Data":"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.947793 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.948946 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8587b466c5-9tcks" event={"ID":"1e19d50d-c0ed-47b7-af00-6e8d2502f599","Type":"ContainerDied","Data":"a4d32343ee4a198199b4fca6f200c4f71160788f8f08f17c4419345d909c1be7"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.948977 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.948992 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5125-account-delete-lrpcc" event={"ID":"baf286a1-0b76-4e66-817c-7d7ce84edaa2","Type":"ContainerDied","Data":"5f4981a54547ee9f0ce3adc448f87d4c7e86f80240ae2ecacd9efac2c2bb4710"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949007 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de","Type":"ContainerDied","Data":"5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949018 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de","Type":"ContainerDied","Data":"e140144ec0a450c4367b9cfe0053b6718e433805a554f25af15235aaeb9f696d"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949027 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e140144ec0a450c4367b9cfe0053b6718e433805a554f25af15235aaeb9f696d" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949035 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0960-account-delete-pc5x5" event={"ID":"572a50eb-42e9-4897-9925-4073b5f6a35d","Type":"ContainerStarted","Data":"69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949059 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949070 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0a5c7-account-delete-j2fhk" event={"ID":"0bc0555a-6c39-4e57-8c30-6a6fc422376a","Type":"ContainerStarted","Data":"0849c2b46e031c79a2d5d154130dc5d8f1d1b025f1f9369edf5dfdd4494048f0"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949080 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0a5c7-account-delete-j2fhk" event={"ID":"0bc0555a-6c39-4e57-8c30-6a6fc422376a","Type":"ContainerStarted","Data":"32802bcbc5b55453efc9309d02e07fb693777d8deb061e6febf04a031e582b91"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949091 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-4sm7k"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949099 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rwzxk"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949110 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rwzxk"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949122 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6b848b8657-gfw2m"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949133 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-4sm7k"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949143 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone8fdd-account-delete-pprlx"] Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949492 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" containerName="mysql-bootstrap" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949502 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" containerName="mysql-bootstrap" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949522 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8260f89c-239d-42e6-b187-c912013d5eca" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949528 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8260f89c-239d-42e6-b187-c912013d5eca" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949537 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df3b0afc-1419-44ce-89f7-b1aca189e145" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949543 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="df3b0afc-1419-44ce-89f7-b1aca189e145" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949836 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="extract-utilities" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949847 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="extract-utilities" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949858 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerName="dnsmasq-dns" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949864 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerName="dnsmasq-dns" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949885 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerName="ovsdbserver-sb" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949892 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerName="ovsdbserver-sb" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949901 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="extract-content" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949907 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="extract-content" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949922 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" containerName="galera" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949928 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" containerName="galera" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.949898 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f69ee496-d325-40e0-81ff-0fb3625a775a" containerName="kube-state-metrics" containerID="cri-o://cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f" gracePeriod=30 Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.949938 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04384280-d303-475d-8abc-c9c957db7fd9" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950026 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="04384280-d303-475d-8abc-c9c957db7fd9" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950027 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-6b848b8657-gfw2m" podUID="e30f3ad7-043b-4212-a496-656c3447b509" containerName="keystone-api" containerID="cri-o://6f451caacded099e931e441c82a111b08382c08a1a8b89ed90e33967d206ea41" gracePeriod=30 Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.950063 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerName="proxy-httpd" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950074 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerName="proxy-httpd" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.950093 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950098 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.950106 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerName="proxy-server" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950111 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerName="proxy-server" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.950121 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="registry-server" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950127 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="registry-server" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.950147 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerName="init" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950154 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerName="init" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:08.950165 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8260f89c-239d-42e6-b187-c912013d5eca" containerName="ovsdbserver-nb" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950171 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8260f89c-239d-42e6-b187-c912013d5eca" containerName="ovsdbserver-nb" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950496 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="ceilometer-central-agent" containerID="cri-o://fd75ea492b2e5d3b5ac427900e214c87df97693bea489cdea1450e4ac5fce395" gracePeriod=30 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950517 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950533 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="df3b0afc-1419-44ce-89f7-b1aca189e145" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950543 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerName="proxy-httpd" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950552 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" containerName="galera" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950559 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" containerName="proxy-server" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950570 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b3d3641-40ca-4391-b7f2-811a85e9c99e" containerName="ovsdbserver-sb" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950584 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8126945c-b3f3-4fc8-a5eb-553c41a439d5" containerName="dnsmasq-dns" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950591 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e3af98a-08a3-4cc1-86f5-9afbf75cc5eb" containerName="registry-server" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950604 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="04384280-d303-475d-8abc-c9c957db7fd9" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950612 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8260f89c-239d-42e6-b187-c912013d5eca" containerName="openstack-network-exporter" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950624 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8260f89c-239d-42e6-b187-c912013d5eca" containerName="ovsdbserver-nb" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950660 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="proxy-httpd" containerID="cri-o://b88d4d8819b799ee0958eb33b8f4190198f52e90c0717a81284a3e8f801ddb08" gracePeriod=30 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950700 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="sg-core" containerID="cri-o://9242e69cc08af9bee9b0d9c36b79f79805314218c4453cc7949a223cbc6ad191" gracePeriod=30 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950787 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="ceilometer-notification-agent" containerID="cri-o://733bbc32986c6be4c2390635eddd73d705ca1fffc3e0db20cc69f0e5434d3516" gracePeriod=30 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.950609 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="b36490b8-d5db-49c3-97e4-b734feaabd00" containerName="memcached" containerID="cri-o://40db7579ead1028473e650e6ae3b6bc768d742221d67f47bd5ce52721da60990" gracePeriod=30 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951236 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone8fdd-account-delete-pprlx"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951253 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951266 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ndzlb"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951277 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ndzlb"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951288 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone8fdd-account-delete-pprlx"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951298 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8fdd-account-create-update-k6bfv"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951308 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8fdd-account-create-update-k6bfv"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951319 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8l256"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951328 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8l256"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951374 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c521-account-create-update-48cv5"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951384 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancec521-account-delete-d62ms"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951394 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c521-account-create-update-48cv5"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951404 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-5f2hw"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951413 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-5f2hw"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951443 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron6e16-account-delete-snvqr"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951453 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6e16-account-create-update-c4t4n"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951463 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6e16-account-create-update-c4t4n"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951471 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-qmrwz"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951480 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-qmrwz"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951489 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement0960-account-delete-pc5x5"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951498 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-0960-account-create-update-2xwzs"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951507 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-0960-account-create-update-2xwzs"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951648 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-fftdj"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951697 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-fftdj"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951727 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone8fdd-account-delete-pprlx" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951797 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-eae6-account-create-update-xrhjp"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951821 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-eae6-account-create-update-xrhjp"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951862 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicaneae6-account-delete-l9467"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951875 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-xj4g5"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951883 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-xj4g5"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951893 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5125-account-create-update-c4zxx"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951902 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5125-account-delete-lrpcc"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:08.951934 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5125-account-create-update-c4zxx"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.002604 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hn6lk"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.020895 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hn6lk"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.027374 4946 generic.go:334] "Generic (PLEG): container finished" podID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerID="7289e3069aa5ae14b8fa7cb3805ddb5abba4f40160f9caf5ebdf388dff4bf65a" exitCode=0 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.027460 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7758850c-e98d-4bb6-bd84-31584a9bcdb6","Type":"ContainerDied","Data":"7289e3069aa5ae14b8fa7cb3805ddb5abba4f40160f9caf5ebdf388dff4bf65a"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.028264 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi2e82-account-delete-77lgb"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.031278 4946 generic.go:334] "Generic (PLEG): container finished" podID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerID="ac9c0edaa246fce85ea02c792d4ac1c07afd3d940a4142fe3a468aaa6e107e16" exitCode=0 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.031325 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69454fc688-t54w7" event={"ID":"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541","Type":"ContainerDied","Data":"ac9c0edaa246fce85ea02c792d4ac1c07afd3d940a4142fe3a468aaa6e107e16"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.038630 4946 generic.go:334] "Generic (PLEG): container finished" podID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerID="6c046b12932defb2b1b37ab4cba5b712c6fb5e6d71e6a3ef814f1586ed8b402d" exitCode=0 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.039066 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bd1d9c96-a0f4-43db-a686-2b320c1948ae","Type":"ContainerDied","Data":"6c046b12932defb2b1b37ab4cba5b712c6fb5e6d71e6a3ef814f1586ed8b402d"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.041273 4946 generic.go:334] "Generic (PLEG): container finished" podID="761f2199-bf20-487a-a18d-172c65a4b515" containerID="364174d07480a501e011d47109a2fdb307a389eee25a760b44b2337057f852ec" exitCode=0 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.041335 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6786b69d4-clggf" event={"ID":"761f2199-bf20-487a-a18d-172c65a4b515","Type":"ContainerDied","Data":"364174d07480a501e011d47109a2fdb307a389eee25a760b44b2337057f852ec"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.044163 4946 generic.go:334] "Generic (PLEG): container finished" podID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerID="434d3710c0181e612e1826bfbdad35d34bd77749c41c0c2a77720e391d7d862e" exitCode=0 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.044520 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df","Type":"ContainerDied","Data":"434d3710c0181e612e1826bfbdad35d34bd77749c41c0c2a77720e391d7d862e"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.044993 4946 scope.go:117] "RemoveContainer" containerID="4aaa4e05b46fa6c68df1bd9487eadf3885468fbce3baff034433d3bc2c281e0b" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.047259 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicaneae6-account-delete-l9467" event={"ID":"85a6dcc4-23cc-4f57-8714-06d0ef25c680","Type":"ContainerStarted","Data":"93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.047622 4946 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbicaneae6-account-delete-l9467" secret="" err="secret \"galera-openstack-dockercfg-nqcgs\" not found" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.055000 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron6e16-account-delete-snvqr" event={"ID":"b6e24ccc-b386-4d6e-9c67-cc95e47cafab","Type":"ContainerStarted","Data":"bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.055260 4946 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron6e16-account-delete-snvqr" secret="" err="secret \"galera-openstack-dockercfg-nqcgs\" not found" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.056095 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.056170 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts podName:85a6dcc4-23cc-4f57-8714-06d0ef25c680 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:09.556150956 +0000 UTC m=+1562.352841065 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts") pod "barbicaneae6-account-delete-l9467" (UID: "85a6dcc4-23cc-4f57-8714-06d0ef25c680") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.063772 4946 generic.go:334] "Generic (PLEG): container finished" podID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerID="99eb604fda1c1f8beeca26feebc6e7d241d1c8fcbc5022252d4d91c35a9e66a0" exitCode=0 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.063838 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"192c2964-d6c9-404e-8cca-cac01c750c5b","Type":"ContainerDied","Data":"99eb604fda1c1f8beeca26feebc6e7d241d1c8fcbc5022252d4d91c35a9e66a0"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.071345 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-2e82-account-create-update-q2kxl"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.076668 4946 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi2e82-account-delete-77lgb" secret="" err="secret \"galera-openstack-dockercfg-nqcgs\" not found" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.077111 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancec521-account-delete-d62ms" event={"ID":"0e57ed09-f7ce-403e-9cb5-676ae1c64117","Type":"ContainerDied","Data":"2e115a720f507036e37b79865616b059c331fb01e03b2871ece60f2c8778b9ee"} Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.077132 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e115a720f507036e37b79865616b059c331fb01e03b2871ece60f2c8778b9ee" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.077170 4946 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placement0960-account-delete-pc5x5" secret="" err="secret \"galera-openstack-dockercfg-nqcgs\" not found" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.095278 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-2e82-account-create-update-q2kxl"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.109900 4946 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell0a5c7-account-delete-j2fhk" secret="" err="secret \"galera-openstack-dockercfg-nqcgs\" not found" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.157767 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.157817 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts podName:b6e24ccc-b386-4d6e-9c67-cc95e47cafab nodeName:}" failed. No retries permitted until 2025-12-03 07:16:09.657802187 +0000 UTC m=+1562.454492296 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts") pod "neutron6e16-account-delete-snvqr" (UID: "b6e24ccc-b386-4d6e-9c67-cc95e47cafab") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.157983 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.158005 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.158012 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts podName:572a50eb-42e9-4897-9925-4073b5f6a35d nodeName:}" failed. No retries permitted until 2025-12-03 07:16:09.658004213 +0000 UTC m=+1562.454694322 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts") pod "placement0960-account-delete-pc5x5" (UID: "572a50eb-42e9-4897-9925-4073b5f6a35d") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.158029 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts podName:2f4c6455-fec3-41e2-bf0a-3595e34d63aa nodeName:}" failed. No retries permitted until 2025-12-03 07:16:09.658022613 +0000 UTC m=+1562.454712722 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts") pod "novaapi2e82-account-delete-77lgb" (UID: "2f4c6455-fec3-41e2-bf0a-3595e34d63aa") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.158036 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.158057 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts podName:0bc0555a-6c39-4e57-8c30-6a6fc422376a nodeName:}" failed. No retries permitted until 2025-12-03 07:16:09.658051014 +0000 UTC m=+1562.454741123 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts") pod "novacell0a5c7-account-delete-j2fhk" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.181956 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-cg9b9"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.210825 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-cg9b9"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.226046 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0a5c7-account-delete-j2fhk"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.232617 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a5c7-account-create-update-qksjg"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.237773 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a5c7-account-create-update-qksjg"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.242454 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbicaneae6-account-delete-l9467" podStartSLOduration=6.242435251 podStartE2EDuration="6.242435251s" podCreationTimestamp="2025-12-03 07:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:16:09.072764112 +0000 UTC m=+1561.869454221" watchObservedRunningTime="2025-12-03 07:16:09.242435251 +0000 UTC m=+1562.039125360" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.244690 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron6e16-account-delete-snvqr" podStartSLOduration=6.244682551 podStartE2EDuration="6.244682551s" podCreationTimestamp="2025-12-03 07:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:16:09.084273811 +0000 UTC m=+1561.880963920" watchObservedRunningTime="2025-12-03 07:16:09.244682551 +0000 UTC m=+1562.041372650" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.307625 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="85593cc4-b6bd-44e9-a908-58f99f0c8756" containerName="galera" containerID="cri-o://a625c10f25213752b4f3ffbe138270f09d8ca2e6105053ba6f35c9e4d430ce6e" gracePeriod=30 Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.311029 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.319624 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.325131 4946 scope.go:117] "RemoveContainer" containerID="b376d6f31b26fb3b5d8428d35df403e2d7ba33125152c97ad1261520a5e2f273" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.354276 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-8587b466c5-9tcks"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.354759 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.357524 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone8fdd-account-delete-pprlx" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360695 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4xmb\" (UniqueName: \"kubernetes.io/projected/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-kube-api-access-h4xmb\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360731 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-logs\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360773 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-etc-machine-id\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360800 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-combined-ca-bundle\") pod \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360823 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-combined-ca-bundle\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360872 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-scripts\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360894 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e57ed09-f7ce-403e-9cb5-676ae1c64117-operator-scripts\") pod \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\" (UID: \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360956 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-internal-tls-certs\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.360994 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-public-tls-certs\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.361027 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-config-data\") pod \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.361057 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g647z\" (UniqueName: \"kubernetes.io/projected/0e57ed09-f7ce-403e-9cb5-676ae1c64117-kube-api-access-g647z\") pod \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\" (UID: \"0e57ed09-f7ce-403e-9cb5-676ae1c64117\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.361089 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.361128 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8drq\" (UniqueName: \"kubernetes.io/projected/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-kube-api-access-k8drq\") pod \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\" (UID: \"10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.361163 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data-custom\") pod \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\" (UID: \"8a3521a3-0f30-4afe-b5f8-efeb6f838e72\") " Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.361187 4946 scope.go:117] "RemoveContainer" containerID="755c77d55195689162b2c7aae3960309c3338cb7b0ded20090fed4b7656c4494" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.364325 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.364390 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-8587b466c5-9tcks"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.365713 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e57ed09-f7ce-403e-9cb5-676ae1c64117-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e57ed09-f7ce-403e-9cb5-676ae1c64117" (UID: "0e57ed09-f7ce-403e-9cb5-676ae1c64117"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.366247 4946 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.366278 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e57ed09-f7ce-403e-9cb5-676ae1c64117-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.371083 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e57ed09-f7ce-403e-9cb5-676ae1c64117-kube-api-access-g647z" (OuterVolumeSpecName: "kube-api-access-g647z") pod "0e57ed09-f7ce-403e-9cb5-676ae1c64117" (UID: "0e57ed09-f7ce-403e-9cb5-676ae1c64117"). InnerVolumeSpecName "kube-api-access-g647z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.372563 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-logs" (OuterVolumeSpecName: "logs") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.374470 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.378942 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.387290 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-scripts" (OuterVolumeSpecName: "scripts") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.391486 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.391700 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-kube-api-access-k8drq" (OuterVolumeSpecName: "kube-api-access-k8drq") pod "10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" (UID: "10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de"). InnerVolumeSpecName "kube-api-access-k8drq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.396920 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.397501 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-kube-api-access-h4xmb" (OuterVolumeSpecName: "kube-api-access-h4xmb") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "kube-api-access-h4xmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.410115 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.429892 4946 scope.go:117] "RemoveContainer" containerID="f019b96d17fb751f8b25510337f077e065922ffd4d5f69477d62724118532891" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.447720 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" (UID: "10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.455196 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.473675 4946 scope.go:117] "RemoveContainer" containerID="ffed1397d4af3d4cd73aff15e45d307e90a8f9c3d60ab75e2f08709cf0802df3" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.476292 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g647z\" (UniqueName: \"kubernetes.io/projected/0e57ed09-f7ce-403e-9cb5-676ae1c64117-kube-api-access-g647z\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.478848 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8drq\" (UniqueName: \"kubernetes.io/projected/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-kube-api-access-k8drq\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.478951 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.479005 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4xmb\" (UniqueName: \"kubernetes.io/projected/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-kube-api-access-h4xmb\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.479053 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.479099 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.479144 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.479190 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.507815 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-config-data" (OuterVolumeSpecName: "config-data") pod "10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" (UID: "10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.513799 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.513826 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.547893 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data" (OuterVolumeSpecName: "config-data") pod "8a3521a3-0f30-4afe-b5f8-efeb6f838e72" (UID: "8a3521a3-0f30-4afe-b5f8-efeb6f838e72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.581137 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.581167 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.581176 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.581185 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3521a3-0f30-4afe-b5f8-efeb6f838e72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.581146 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.581243 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts podName:85a6dcc4-23cc-4f57-8714-06d0ef25c680 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:10.581226485 +0000 UTC m=+1563.377916594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts") pod "barbicaneae6-account-delete-l9467" (UID: "85a6dcc4-23cc-4f57-8714-06d0ef25c680") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.605525 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04384280-d303-475d-8abc-c9c957db7fd9" path="/var/lib/kubelet/pods/04384280-d303-475d-8abc-c9c957db7fd9/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.606328 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17d6a59c-9fc5-47c3-b296-a32eec387ade" path="/var/lib/kubelet/pods/17d6a59c-9fc5-47c3-b296-a32eec387ade/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.606884 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e19d50d-c0ed-47b7-af00-6e8d2502f599" path="/var/lib/kubelet/pods/1e19d50d-c0ed-47b7-af00-6e8d2502f599/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.609419 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e323f7f-74ef-43bc-a54a-a061675e3dbf" path="/var/lib/kubelet/pods/2e323f7f-74ef-43bc-a54a-a061675e3dbf/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.610244 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b4bacb1-58ed-42a2-a98f-b871b1f1047f" path="/var/lib/kubelet/pods/3b4bacb1-58ed-42a2-a98f-b871b1f1047f/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.610721 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c6a0a88-de49-4843-b241-1c0d25c6230a" path="/var/lib/kubelet/pods/3c6a0a88-de49-4843-b241-1c0d25c6230a/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.611240 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cb069b9-d659-481c-bbb5-616c4af71d65" path="/var/lib/kubelet/pods/4cb069b9-d659-481c-bbb5-616c4af71d65/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.612241 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ebfd4c9-aa01-4ab5-82d7-506d0b97b573" path="/var/lib/kubelet/pods/5ebfd4c9-aa01-4ab5-82d7-506d0b97b573/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.612754 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79eeefc8-7a29-424f-b13f-40b859f7ece0" path="/var/lib/kubelet/pods/79eeefc8-7a29-424f-b13f-40b859f7ece0/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.613230 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a18e489-bc4f-462a-87da-280408bb99e2" path="/var/lib/kubelet/pods/7a18e489-bc4f-462a-87da-280408bb99e2/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.615837 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e289880-83eb-42c7-adbe-319368792548" path="/var/lib/kubelet/pods/7e289880-83eb-42c7-adbe-319368792548/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.616774 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd" path="/var/lib/kubelet/pods/7e8e38b6-90cf-4c68-ac4e-06b2ee1e30cd/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.617587 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8489b683-060b-438e-a8eb-0615a536c35d" path="/var/lib/kubelet/pods/8489b683-060b-438e-a8eb-0615a536c35d/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.618864 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="880093ae-378a-4b62-902b-35b216d6d022" path="/var/lib/kubelet/pods/880093ae-378a-4b62-902b-35b216d6d022/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.619728 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9572139a-335a-417b-a498-66408e40ae18" path="/var/lib/kubelet/pods/9572139a-335a-417b-a498-66408e40ae18/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.620630 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eeb90cf-935a-4792-a862-31ba108d49c8" path="/var/lib/kubelet/pods/9eeb90cf-935a-4792-a862-31ba108d49c8/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.621259 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b7bfe9-8f09-44f1-8966-3c1df805df6c" path="/var/lib/kubelet/pods/a0b7bfe9-8f09-44f1-8966-3c1df805df6c/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.622155 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdc5e484-c93b-47bf-8144-e76a5db68bfd" path="/var/lib/kubelet/pods/cdc5e484-c93b-47bf-8144-e76a5db68bfd/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.623157 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dff5e6e2-1e97-4aef-8cde-be3c301cb763" path="/var/lib/kubelet/pods/dff5e6e2-1e97-4aef-8cde-be3c301cb763/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.623920 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e282365b-8ff7-4fd8-9220-ccdac9865543" path="/var/lib/kubelet/pods/e282365b-8ff7-4fd8-9220-ccdac9865543/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.624961 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea1b7f5d-0115-4bd2-8d9c-502eff1ae345" path="/var/lib/kubelet/pods/ea1b7f5d-0115-4bd2-8d9c-502eff1ae345/volumes" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.658586 4946 scope.go:117] "RemoveContainer" containerID="c1b84099e3ff7baff4b0d71a61709731ec4d0a9f91be8cac1668b867cdc1f2bf" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.689270 4946 scope.go:117] "RemoveContainer" containerID="73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.692078 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.692153 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts podName:2f4c6455-fec3-41e2-bf0a-3595e34d63aa nodeName:}" failed. No retries permitted until 2025-12-03 07:16:10.692131796 +0000 UTC m=+1563.488821905 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts") pod "novaapi2e82-account-delete-77lgb" (UID: "2f4c6455-fec3-41e2-bf0a-3595e34d63aa") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.692644 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.692698 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts podName:b6e24ccc-b386-4d6e-9c67-cc95e47cafab nodeName:}" failed. No retries permitted until 2025-12-03 07:16:10.692685961 +0000 UTC m=+1563.489376070 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts") pod "neutron6e16-account-delete-snvqr" (UID: "b6e24ccc-b386-4d6e-9c67-cc95e47cafab") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.692808 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.692855 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts podName:572a50eb-42e9-4897-9925-4073b5f6a35d nodeName:}" failed. No retries permitted until 2025-12-03 07:16:10.692844575 +0000 UTC m=+1563.489534694 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts") pod "placement0960-account-delete-pc5x5" (UID: "572a50eb-42e9-4897-9925-4073b5f6a35d") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.693004 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.693118 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts podName:0bc0555a-6c39-4e57-8c30-6a6fc422376a nodeName:}" failed. No retries permitted until 2025-12-03 07:16:10.693095882 +0000 UTC m=+1563.489785991 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts") pod "novacell0a5c7-account-delete-j2fhk" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a") : configmap "openstack-scripts" not found Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.744640 4946 scope.go:117] "RemoveContainer" containerID="2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.782217 4946 scope.go:117] "RemoveContainer" containerID="73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.784131 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8\": container with ID starting with 73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8 not found: ID does not exist" containerID="73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.784158 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8"} err="failed to get container status \"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8\": rpc error: code = NotFound desc = could not find container \"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8\": container with ID starting with 73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8 not found: ID does not exist" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.784179 4946 scope.go:117] "RemoveContainer" containerID="2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.789136 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1\": container with ID starting with 2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1 not found: ID does not exist" containerID="2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.789159 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1"} err="failed to get container status \"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1\": rpc error: code = NotFound desc = could not find container \"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1\": container with ID starting with 2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1 not found: ID does not exist" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.789176 4946 scope.go:117] "RemoveContainer" containerID="73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.791426 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8"} err="failed to get container status \"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8\": rpc error: code = NotFound desc = could not find container \"73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8\": container with ID starting with 73d42fa7af6bd2b058ba089c88489c90a9c34fc350575718459f2461f41d06b8 not found: ID does not exist" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.791459 4946 scope.go:117] "RemoveContainer" containerID="2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.797077 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1"} err="failed to get container status \"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1\": rpc error: code = NotFound desc = could not find container \"2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1\": container with ID starting with 2b898d59519a274d4f710cd0a700509ba95ed4d8aeb6966732a3a9910ff5eed1 not found: ID does not exist" Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.895067 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217 is running failed: container process not found" containerID="d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.895937 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217 is running failed: container process not found" containerID="d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.898391 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217 is running failed: container process not found" containerID="d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 03 07:16:09 crc kubenswrapper[4946]: E1203 07:16:09.898438 4946 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="0e0b4114-0121-48cc-9439-bc5d350aeb44" containerName="nova-cell0-conductor-conductor" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.909714 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.918133 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.922717 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.944606 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.949303 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.960543 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:16:09 crc kubenswrapper[4946]: I1203 07:16:09.992898 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002497 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-httpd-run\") pod \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002538 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-config-data\") pod \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002567 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-internal-tls-certs\") pod \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002588 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-public-tls-certs\") pod \"761f2199-bf20-487a-a18d-172c65a4b515\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002612 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-combined-ca-bundle\") pod \"0e0b4114-0121-48cc-9439-bc5d350aeb44\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002705 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-internal-tls-certs\") pod \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002728 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-config-data\") pod \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002763 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002805 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002829 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-combined-ca-bundle\") pod \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002851 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-combined-ca-bundle\") pod \"761f2199-bf20-487a-a18d-172c65a4b515\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002891 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjx9b\" (UniqueName: \"kubernetes.io/projected/bd1d9c96-a0f4-43db-a686-2b320c1948ae-kube-api-access-wjx9b\") pod \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002938 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvnbh\" (UniqueName: \"kubernetes.io/projected/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-kube-api-access-zvnbh\") pod \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.002989 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-logs\") pod \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003047 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmwrc\" (UniqueName: \"kubernetes.io/projected/761f2199-bf20-487a-a18d-172c65a4b515-kube-api-access-xmwrc\") pod \"761f2199-bf20-487a-a18d-172c65a4b515\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003082 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-config-data\") pod \"192c2964-d6c9-404e-8cca-cac01c750c5b\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003105 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-scripts\") pod \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003124 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data-custom\") pod \"761f2199-bf20-487a-a18d-172c65a4b515\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003158 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/761f2199-bf20-487a-a18d-172c65a4b515-logs\") pod \"761f2199-bf20-487a-a18d-172c65a4b515\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003174 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data\") pod \"761f2199-bf20-487a-a18d-172c65a4b515\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003195 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-logs\") pod \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003213 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-combined-ca-bundle\") pod \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003235 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwtk2\" (UniqueName: \"kubernetes.io/projected/0e0b4114-0121-48cc-9439-bc5d350aeb44-kube-api-access-fwtk2\") pod \"0e0b4114-0121-48cc-9439-bc5d350aeb44\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003253 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp7fn\" (UniqueName: \"kubernetes.io/projected/192c2964-d6c9-404e-8cca-cac01c750c5b-kube-api-access-qp7fn\") pod \"192c2964-d6c9-404e-8cca-cac01c750c5b\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003269 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-combined-ca-bundle\") pod \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003285 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/192c2964-d6c9-404e-8cca-cac01c750c5b-logs\") pod \"192c2964-d6c9-404e-8cca-cac01c750c5b\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003329 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-config-data\") pod \"0e0b4114-0121-48cc-9439-bc5d350aeb44\" (UID: \"0e0b4114-0121-48cc-9439-bc5d350aeb44\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003345 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-internal-tls-certs\") pod \"761f2199-bf20-487a-a18d-172c65a4b515\" (UID: \"761f2199-bf20-487a-a18d-172c65a4b515\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003374 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-logs\") pod \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003389 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-nova-metadata-tls-certs\") pod \"192c2964-d6c9-404e-8cca-cac01c750c5b\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003405 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-public-tls-certs\") pod \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003423 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-combined-ca-bundle\") pod \"192c2964-d6c9-404e-8cca-cac01c750c5b\" (UID: \"192c2964-d6c9-404e-8cca-cac01c750c5b\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003448 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-scripts\") pod \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003474 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-public-tls-certs\") pod \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\" (UID: \"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003492 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-httpd-run\") pod \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003511 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-config-data\") pod \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\" (UID: \"bd1d9c96-a0f4-43db-a686-2b320c1948ae\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003527 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6mwj\" (UniqueName: \"kubernetes.io/projected/7758850c-e98d-4bb6-bd84-31584a9bcdb6-kube-api-access-t6mwj\") pod \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\" (UID: \"7758850c-e98d-4bb6-bd84-31584a9bcdb6\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.003990 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.011347 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bd1d9c96-a0f4-43db-a686-2b320c1948ae" (UID: "bd1d9c96-a0f4-43db-a686-2b320c1948ae"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.011466 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-logs" (OuterVolumeSpecName: "logs") pod "bd1d9c96-a0f4-43db-a686-2b320c1948ae" (UID: "bd1d9c96-a0f4-43db-a686-2b320c1948ae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.012086 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7758850c-e98d-4bb6-bd84-31584a9bcdb6" (UID: "7758850c-e98d-4bb6-bd84-31584a9bcdb6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.013962 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-logs" (OuterVolumeSpecName: "logs") pod "7758850c-e98d-4bb6-bd84-31584a9bcdb6" (UID: "7758850c-e98d-4bb6-bd84-31584a9bcdb6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.014071 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.015636 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-logs" (OuterVolumeSpecName: "logs") pod "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" (UID: "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.015783 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/761f2199-bf20-487a-a18d-172c65a4b515-logs" (OuterVolumeSpecName: "logs") pod "761f2199-bf20-487a-a18d-172c65a4b515" (UID: "761f2199-bf20-487a-a18d-172c65a4b515"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.017521 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/192c2964-d6c9-404e-8cca-cac01c750c5b-logs" (OuterVolumeSpecName: "logs") pod "192c2964-d6c9-404e-8cca-cac01c750c5b" (UID: "192c2964-d6c9-404e-8cca-cac01c750c5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.018091 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/761f2199-bf20-487a-a18d-172c65a4b515-kube-api-access-xmwrc" (OuterVolumeSpecName: "kube-api-access-xmwrc") pod "761f2199-bf20-487a-a18d-172c65a4b515" (UID: "761f2199-bf20-487a-a18d-172c65a4b515"). InnerVolumeSpecName "kube-api-access-xmwrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.024622 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "7758850c-e98d-4bb6-bd84-31584a9bcdb6" (UID: "7758850c-e98d-4bb6-bd84-31584a9bcdb6"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.038965 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7758850c-e98d-4bb6-bd84-31584a9bcdb6-kube-api-access-t6mwj" (OuterVolumeSpecName: "kube-api-access-t6mwj") pod "7758850c-e98d-4bb6-bd84-31584a9bcdb6" (UID: "7758850c-e98d-4bb6-bd84-31584a9bcdb6"). InnerVolumeSpecName "kube-api-access-t6mwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.039183 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "bd1d9c96-a0f4-43db-a686-2b320c1948ae" (UID: "bd1d9c96-a0f4-43db-a686-2b320c1948ae"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.039266 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "761f2199-bf20-487a-a18d-172c65a4b515" (UID: "761f2199-bf20-487a-a18d-172c65a4b515"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.039282 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-scripts" (OuterVolumeSpecName: "scripts") pod "7758850c-e98d-4bb6-bd84-31584a9bcdb6" (UID: "7758850c-e98d-4bb6-bd84-31584a9bcdb6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.039545 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd1d9c96-a0f4-43db-a686-2b320c1948ae-kube-api-access-wjx9b" (OuterVolumeSpecName: "kube-api-access-wjx9b") pod "bd1d9c96-a0f4-43db-a686-2b320c1948ae" (UID: "bd1d9c96-a0f4-43db-a686-2b320c1948ae"). InnerVolumeSpecName "kube-api-access-wjx9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.041955 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69454fc688-t54w7" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.046647 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-kube-api-access-zvnbh" (OuterVolumeSpecName: "kube-api-access-zvnbh") pod "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" (UID: "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df"). InnerVolumeSpecName "kube-api-access-zvnbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.052862 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-scripts" (OuterVolumeSpecName: "scripts") pod "bd1d9c96-a0f4-43db-a686-2b320c1948ae" (UID: "bd1d9c96-a0f4-43db-a686-2b320c1948ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.052907 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/192c2964-d6c9-404e-8cca-cac01c750c5b-kube-api-access-qp7fn" (OuterVolumeSpecName: "kube-api-access-qp7fn") pod "192c2964-d6c9-404e-8cca-cac01c750c5b" (UID: "192c2964-d6c9-404e-8cca-cac01c750c5b"). InnerVolumeSpecName "kube-api-access-qp7fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.058891 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e0b4114-0121-48cc-9439-bc5d350aeb44-kube-api-access-fwtk2" (OuterVolumeSpecName: "kube-api-access-fwtk2") pod "0e0b4114-0121-48cc-9439-bc5d350aeb44" (UID: "0e0b4114-0121-48cc-9439-bc5d350aeb44"). InnerVolumeSpecName "kube-api-access-fwtk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.098360 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" (UID: "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105194 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-config\") pod \"f69ee496-d325-40e0-81ff-0fb3625a775a\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105263 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmxsp\" (UniqueName: \"kubernetes.io/projected/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-api-access-fmxsp\") pod \"f69ee496-d325-40e0-81ff-0fb3625a775a\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105369 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-combined-ca-bundle\") pod \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105425 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-logs\") pod \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105568 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-config-data\") pod \"e5415d2e-cf73-4dd8-900d-211b957e829f\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105654 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwmjt\" (UniqueName: \"kubernetes.io/projected/e5415d2e-cf73-4dd8-900d-211b957e829f-kube-api-access-wwmjt\") pod \"e5415d2e-cf73-4dd8-900d-211b957e829f\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105679 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46dbr\" (UniqueName: \"kubernetes.io/projected/baf286a1-0b76-4e66-817c-7d7ce84edaa2-kube-api-access-46dbr\") pod \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\" (UID: \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105806 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-certs\") pod \"f69ee496-d325-40e0-81ff-0fb3625a775a\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105826 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-public-tls-certs\") pod \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105852 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-config-data\") pod \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105888 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-internal-tls-certs\") pod \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105903 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-combined-ca-bundle\") pod \"e5415d2e-cf73-4dd8-900d-211b957e829f\" (UID: \"e5415d2e-cf73-4dd8-900d-211b957e829f\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.105952 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp6zn\" (UniqueName: \"kubernetes.io/projected/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-kube-api-access-zp6zn\") pod \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.106082 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf286a1-0b76-4e66-817c-7d7ce84edaa2-operator-scripts\") pod \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\" (UID: \"baf286a1-0b76-4e66-817c-7d7ce84edaa2\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.106113 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-combined-ca-bundle\") pod \"f69ee496-d325-40e0-81ff-0fb3625a775a\" (UID: \"f69ee496-d325-40e0-81ff-0fb3625a775a\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.106145 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-scripts\") pod \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\" (UID: \"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107668 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjx9b\" (UniqueName: \"kubernetes.io/projected/bd1d9c96-a0f4-43db-a686-2b320c1948ae-kube-api-access-wjx9b\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107693 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvnbh\" (UniqueName: \"kubernetes.io/projected/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-kube-api-access-zvnbh\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107704 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107716 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmwrc\" (UniqueName: \"kubernetes.io/projected/761f2199-bf20-487a-a18d-172c65a4b515-kube-api-access-xmwrc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107728 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107759 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107790 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/761f2199-bf20-487a-a18d-172c65a4b515-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107800 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107810 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107821 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwtk2\" (UniqueName: \"kubernetes.io/projected/0e0b4114-0121-48cc-9439-bc5d350aeb44-kube-api-access-fwtk2\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107831 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp7fn\" (UniqueName: \"kubernetes.io/projected/192c2964-d6c9-404e-8cca-cac01c750c5b-kube-api-access-qp7fn\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107842 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/192c2964-d6c9-404e-8cca-cac01c750c5b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107851 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107861 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107870 4946 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7758850c-e98d-4bb6-bd84-31584a9bcdb6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107880 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6mwj\" (UniqueName: \"kubernetes.io/projected/7758850c-e98d-4bb6-bd84-31584a9bcdb6-kube-api-access-t6mwj\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.108159 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-logs" (OuterVolumeSpecName: "logs") pod "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" (UID: "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.107890 4946 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd1d9c96-a0f4-43db-a686-2b320c1948ae-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.108528 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.108542 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.108801 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baf286a1-0b76-4e66-817c-7d7ce84edaa2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "baf286a1-0b76-4e66-817c-7d7ce84edaa2" (UID: "baf286a1-0b76-4e66-817c-7d7ce84edaa2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.120409 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-scripts" (OuterVolumeSpecName: "scripts") pod "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" (UID: "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.122593 4946 generic.go:334] "Generic (PLEG): container finished" podID="f69ee496-d325-40e0-81ff-0fb3625a775a" containerID="cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f" exitCode=2 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.122659 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f69ee496-d325-40e0-81ff-0fb3625a775a","Type":"ContainerDied","Data":"cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.122685 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f69ee496-d325-40e0-81ff-0fb3625a775a","Type":"ContainerDied","Data":"961d002cc0d977266c82641602e07919bb97e9be177176ffddc7cd5b172c7e24"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.122705 4946 scope.go:117] "RemoveContainer" containerID="cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.122814 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.123471 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e0b4114-0121-48cc-9439-bc5d350aeb44" (UID: "0e0b4114-0121-48cc-9439-bc5d350aeb44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.128761 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-kube-api-access-zp6zn" (OuterVolumeSpecName: "kube-api-access-zp6zn") pod "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" (UID: "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541"). InnerVolumeSpecName "kube-api-access-zp6zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.128767 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69454fc688-t54w7" event={"ID":"ce3ead5e-98d3-4719-b1d2-cda4d7b7b541","Type":"ContainerDied","Data":"f0dbde89ae62dd0d52329bd682bf4a8a258dd6f7c276606d012df0778913d590"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.128830 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69454fc688-t54w7" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.129936 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5415d2e-cf73-4dd8-900d-211b957e829f-kube-api-access-wwmjt" (OuterVolumeSpecName: "kube-api-access-wwmjt") pod "e5415d2e-cf73-4dd8-900d-211b957e829f" (UID: "e5415d2e-cf73-4dd8-900d-211b957e829f"). InnerVolumeSpecName "kube-api-access-wwmjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.132589 4946 generic.go:334] "Generic (PLEG): container finished" podID="0e0b4114-0121-48cc-9439-bc5d350aeb44" containerID="d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217" exitCode=0 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.132674 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.132679 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0e0b4114-0121-48cc-9439-bc5d350aeb44","Type":"ContainerDied","Data":"d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.132732 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0e0b4114-0121-48cc-9439-bc5d350aeb44","Type":"ContainerDied","Data":"2a7a80c3ff2375af3de6f56549364816f5acd049a27afae2840afd3d289e3190"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.136124 4946 generic.go:334] "Generic (PLEG): container finished" podID="b36490b8-d5db-49c3-97e4-b734feaabd00" containerID="40db7579ead1028473e650e6ae3b6bc768d742221d67f47bd5ce52721da60990" exitCode=0 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.136193 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b36490b8-d5db-49c3-97e4-b734feaabd00","Type":"ContainerDied","Data":"40db7579ead1028473e650e6ae3b6bc768d742221d67f47bd5ce52721da60990"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.138553 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7758850c-e98d-4bb6-bd84-31584a9bcdb6","Type":"ContainerDied","Data":"925234f7444f86715154173321a353e7552589fab9828c5f28ecc5539d42fb86"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.138621 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.139922 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "761f2199-bf20-487a-a18d-172c65a4b515" (UID: "761f2199-bf20-487a-a18d-172c65a4b515"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.141610 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5125-account-delete-lrpcc" event={"ID":"baf286a1-0b76-4e66-817c-7d7ce84edaa2","Type":"ContainerDied","Data":"03ef8c2e844685ca4de40a06f36b963e777a38ea495584831ab51b98dd3d0a77"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.142828 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03ef8c2e844685ca4de40a06f36b963e777a38ea495584831ab51b98dd3d0a77" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.143627 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.143713 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"192c2964-d6c9-404e-8cca-cac01c750c5b","Type":"ContainerDied","Data":"f052ad62e4e6729b8aeeb5ac686c3fd61a7a04841e36e2178b8a4744bbe0b9fe"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.145070 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5125-account-delete-lrpcc" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.147353 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bd1d9c96-a0f4-43db-a686-2b320c1948ae","Type":"ContainerDied","Data":"3dce7a4effecb50c6bfddb3992c650827125bd582d3634d4d3e8261ace1d7b4c"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.147357 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.150841 4946 scope.go:117] "RemoveContainer" containerID="cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.154871 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6786b69d4-clggf" event={"ID":"761f2199-bf20-487a-a18d-172c65a4b515","Type":"ContainerDied","Data":"bec07d9cc9516495d2dc8ed1b4d9325ba551821764e91c5734a0a55d91fa7ed5"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.154973 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6786b69d4-clggf" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.160040 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-config-data" (OuterVolumeSpecName: "config-data") pod "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" (UID: "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.160494 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "761f2199-bf20-487a-a18d-172c65a4b515" (UID: "761f2199-bf20-487a-a18d-172c65a4b515"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.160550 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-api-access-fmxsp" (OuterVolumeSpecName: "kube-api-access-fmxsp") pod "f69ee496-d325-40e0-81ff-0fb3625a775a" (UID: "f69ee496-d325-40e0-81ff-0fb3625a775a"). InnerVolumeSpecName "kube-api-access-fmxsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.162201 4946 generic.go:334] "Generic (PLEG): container finished" podID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerID="b88d4d8819b799ee0958eb33b8f4190198f52e90c0717a81284a3e8f801ddb08" exitCode=0 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.162237 4946 generic.go:334] "Generic (PLEG): container finished" podID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerID="9242e69cc08af9bee9b0d9c36b79f79805314218c4453cc7949a223cbc6ad191" exitCode=2 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.162247 4946 generic.go:334] "Generic (PLEG): container finished" podID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerID="fd75ea492b2e5d3b5ac427900e214c87df97693bea489cdea1450e4ac5fce395" exitCode=0 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.162308 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerDied","Data":"b88d4d8819b799ee0958eb33b8f4190198f52e90c0717a81284a3e8f801ddb08"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.162346 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerDied","Data":"9242e69cc08af9bee9b0d9c36b79f79805314218c4453cc7949a223cbc6ad191"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.162356 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerDied","Data":"fd75ea492b2e5d3b5ac427900e214c87df97693bea489cdea1450e4ac5fce395"} Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.162382 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f\": container with ID starting with cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f not found: ID does not exist" containerID="cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.162405 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f"} err="failed to get container status \"cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f\": rpc error: code = NotFound desc = could not find container \"cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f\": container with ID starting with cc2350979bf2fa0fe53ec1a63e17e9792b4cf11931b0256995ea508813f76b8f not found: ID does not exist" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.162424 4946 scope.go:117] "RemoveContainer" containerID="ac9c0edaa246fce85ea02c792d4ac1c07afd3d940a4142fe3a468aaa6e107e16" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.167325 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4dc4eff7-f011-4265-b9dd-4e6bcdcb08df","Type":"ContainerDied","Data":"47ddc59169e2f57c7bc4a899a86665600d584abf00eecf20b360926ed8b20fb4"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.167401 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.171977 4946 generic.go:334] "Generic (PLEG): container finished" podID="e5415d2e-cf73-4dd8-900d-211b957e829f" containerID="5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6" exitCode=0 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.172056 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5415d2e-cf73-4dd8-900d-211b957e829f","Type":"ContainerDied","Data":"5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.172083 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5415d2e-cf73-4dd8-900d-211b957e829f","Type":"ContainerDied","Data":"db27503cb03cb197511a3037f23064a2773acab8bcb40b49a1d67ee7d47f7e26"} Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.172139 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.177557 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.180201 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-config-data" (OuterVolumeSpecName: "config-data") pod "192c2964-d6c9-404e-8cca-cac01c750c5b" (UID: "192c2964-d6c9-404e-8cca-cac01c750c5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.181602 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.181980 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baf286a1-0b76-4e66-817c-7d7ce84edaa2-kube-api-access-46dbr" (OuterVolumeSpecName: "kube-api-access-46dbr") pod "baf286a1-0b76-4e66-817c-7d7ce84edaa2" (UID: "baf286a1-0b76-4e66-817c-7d7ce84edaa2"). InnerVolumeSpecName "kube-api-access-46dbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.182149 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapi2e82-account-delete-77lgb" podUID="2f4c6455-fec3-41e2-bf0a-3595e34d63aa" containerName="mariadb-account-delete" containerID="cri-o://11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d" gracePeriod=30 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.182227 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.182160 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancec521-account-delete-d62ms" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.182676 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone8fdd-account-delete-pprlx" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.182805 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbicaneae6-account-delete-l9467" podUID="85a6dcc4-23cc-4f57-8714-06d0ef25c680" containerName="mariadb-account-delete" containerID="cri-o://93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932" gracePeriod=30 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.183024 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron6e16-account-delete-snvqr" podUID="b6e24ccc-b386-4d6e-9c67-cc95e47cafab" containerName="mariadb-account-delete" containerID="cri-o://bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b" gracePeriod=30 Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.183151 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell0a5c7-account-delete-j2fhk" podUID="0bc0555a-6c39-4e57-8c30-6a6fc422376a" containerName="mariadb-account-delete" containerID="cri-o://0849c2b46e031c79a2d5d154130dc5d8f1d1b025f1f9369edf5dfdd4494048f0" gracePeriod=30 Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.187032 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.187235 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement0960-account-delete-pc5x5" podUID="572a50eb-42e9-4897-9925-4073b5f6a35d" containerName="mariadb-account-delete" containerID="cri-o://69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526" gracePeriod=30 Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.190029 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.190865 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.190893 4946 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.200634 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.204039 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.206253 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.206319 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.206663 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" (UID: "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211077 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmxsp\" (UniqueName: \"kubernetes.io/projected/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-api-access-fmxsp\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211111 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211125 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211138 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211151 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211163 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211174 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211185 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwmjt\" (UniqueName: \"kubernetes.io/projected/e5415d2e-cf73-4dd8-900d-211b957e829f-kube-api-access-wwmjt\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211195 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46dbr\" (UniqueName: \"kubernetes.io/projected/baf286a1-0b76-4e66-817c-7d7ce84edaa2-kube-api-access-46dbr\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211205 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211216 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp6zn\" (UniqueName: \"kubernetes.io/projected/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-kube-api-access-zp6zn\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211228 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baf286a1-0b76-4e66-817c-7d7ce84edaa2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211239 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.211251 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.213075 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-config-data" (OuterVolumeSpecName: "config-data") pod "e5415d2e-cf73-4dd8-900d-211b957e829f" (UID: "e5415d2e-cf73-4dd8-900d-211b957e829f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.213497 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7758850c-e98d-4bb6-bd84-31584a9bcdb6" (UID: "7758850c-e98d-4bb6-bd84-31584a9bcdb6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.238247 4946 scope.go:117] "RemoveContainer" containerID="6a378fe14ccec4bc8e63806ae734f9293c3f59cfa0f3b3d5fb967e83eb1a42a1" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.243351 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" (UID: "4dc4eff7-f011-4265-b9dd-4e6bcdcb08df"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.245608 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.254809 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "f69ee496-d325-40e0-81ff-0fb3625a775a" (UID: "f69ee496-d325-40e0-81ff-0fb3625a775a"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.258291 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.281041 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd1d9c96-a0f4-43db-a686-2b320c1948ae" (UID: "bd1d9c96-a0f4-43db-a686-2b320c1948ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.299648 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.313026 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.314584 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.314622 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.314635 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.314648 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.314662 4946 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.319238 4946 scope.go:117] "RemoveContainer" containerID="d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.320477 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.332969 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone8fdd-account-delete-pprlx"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.340905 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone8fdd-account-delete-pprlx"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.343452 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" (UID: "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.346384 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-config-data" (OuterVolumeSpecName: "config-data") pod "0e0b4114-0121-48cc-9439-bc5d350aeb44" (UID: "0e0b4114-0121-48cc-9439-bc5d350aeb44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.348659 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancec521-account-delete-d62ms"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.352303 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "192c2964-d6c9-404e-8cca-cac01c750c5b" (UID: "192c2964-d6c9-404e-8cca-cac01c750c5b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.352387 4946 scope.go:117] "RemoveContainer" containerID="d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217" Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.353761 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217\": container with ID starting with d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217 not found: ID does not exist" containerID="d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.353798 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217"} err="failed to get container status \"d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217\": rpc error: code = NotFound desc = could not find container \"d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217\": container with ID starting with d99bbb27e06c29bcd5c53518197d17ed1ddbb3c8842c5e697a520eb079453217 not found: ID does not exist" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.353819 4946 scope.go:117] "RemoveContainer" containerID="7289e3069aa5ae14b8fa7cb3805ddb5abba4f40160f9caf5ebdf388dff4bf65a" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.355770 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancec521-account-delete-d62ms"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.364591 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-config-data" (OuterVolumeSpecName: "config-data") pod "7758850c-e98d-4bb6-bd84-31584a9bcdb6" (UID: "7758850c-e98d-4bb6-bd84-31584a9bcdb6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.375752 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-config-data" (OuterVolumeSpecName: "config-data") pod "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" (UID: "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.385390 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5415d2e-cf73-4dd8-900d-211b957e829f" (UID: "e5415d2e-cf73-4dd8-900d-211b957e829f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.385472 4946 scope.go:117] "RemoveContainer" containerID="33b47ed2f51fb1905563dad28b7eecdde497b32b4bd342ccb3247ecc1b377e2f" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.396032 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bd1d9c96-a0f4-43db-a686-2b320c1948ae" (UID: "bd1d9c96-a0f4-43db-a686-2b320c1948ae"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.405218 4946 scope.go:117] "RemoveContainer" containerID="99eb604fda1c1f8beeca26feebc6e7d241d1c8fcbc5022252d4d91c35a9e66a0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.413583 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "f69ee496-d325-40e0-81ff-0fb3625a775a" (UID: "f69ee496-d325-40e0-81ff-0fb3625a775a"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416125 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416335 4946 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416368 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416383 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5415d2e-cf73-4dd8-900d-211b957e829f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416395 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e0b4114-0121-48cc-9439-bc5d350aeb44-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416407 4946 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416492 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416545 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416558 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.416572 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.465845 4946 scope.go:117] "RemoveContainer" containerID="04a7a8335ee4dd2cc59886e31042cdc9337381a24ca52a361f5d5f35b6bdead5" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.466029 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "192c2964-d6c9-404e-8cca-cac01c750c5b" (UID: "192c2964-d6c9-404e-8cca-cac01c750c5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.480983 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f69ee496-d325-40e0-81ff-0fb3625a775a" (UID: "f69ee496-d325-40e0-81ff-0fb3625a775a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.482344 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7758850c-e98d-4bb6-bd84-31584a9bcdb6" (UID: "7758850c-e98d-4bb6-bd84-31584a9bcdb6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.488365 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "761f2199-bf20-487a-a18d-172c65a4b515" (UID: "761f2199-bf20-487a-a18d-172c65a4b515"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.494515 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data" (OuterVolumeSpecName: "config-data") pod "761f2199-bf20-487a-a18d-172c65a4b515" (UID: "761f2199-bf20-487a-a18d-172c65a4b515"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.503380 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-config-data" (OuterVolumeSpecName: "config-data") pod "bd1d9c96-a0f4-43db-a686-2b320c1948ae" (UID: "bd1d9c96-a0f4-43db-a686-2b320c1948ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.517105 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-combined-ca-bundle\") pod \"b36490b8-d5db-49c3-97e4-b734feaabd00\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.517219 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-config-data\") pod \"b36490b8-d5db-49c3-97e4-b734feaabd00\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.517311 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2k7l\" (UniqueName: \"kubernetes.io/projected/b36490b8-d5db-49c3-97e4-b734feaabd00-kube-api-access-h2k7l\") pod \"b36490b8-d5db-49c3-97e4-b734feaabd00\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.517345 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-memcached-tls-certs\") pod \"b36490b8-d5db-49c3-97e4-b734feaabd00\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.517429 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-kolla-config\") pod \"b36490b8-d5db-49c3-97e4-b734feaabd00\" (UID: \"b36490b8-d5db-49c3-97e4-b734feaabd00\") " Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.518066 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd1d9c96-a0f4-43db-a686-2b320c1948ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.518092 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.518106 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7758850c-e98d-4bb6-bd84-31584a9bcdb6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.518119 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761f2199-bf20-487a-a18d-172c65a4b515-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.518130 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ee496-d325-40e0-81ff-0fb3625a775a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.518141 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/192c2964-d6c9-404e-8cca-cac01c750c5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.521498 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36490b8-d5db-49c3-97e4-b734feaabd00-kube-api-access-h2k7l" (OuterVolumeSpecName: "kube-api-access-h2k7l") pod "b36490b8-d5db-49c3-97e4-b734feaabd00" (UID: "b36490b8-d5db-49c3-97e4-b734feaabd00"). InnerVolumeSpecName "kube-api-access-h2k7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.521937 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-config-data" (OuterVolumeSpecName: "config-data") pod "b36490b8-d5db-49c3-97e4-b734feaabd00" (UID: "b36490b8-d5db-49c3-97e4-b734feaabd00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.524158 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "b36490b8-d5db-49c3-97e4-b734feaabd00" (UID: "b36490b8-d5db-49c3-97e4-b734feaabd00"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.545948 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b36490b8-d5db-49c3-97e4-b734feaabd00" (UID: "b36490b8-d5db-49c3-97e4-b734feaabd00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.579558 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" (UID: "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.589824 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "b36490b8-d5db-49c3-97e4-b734feaabd00" (UID: "b36490b8-d5db-49c3-97e4-b734feaabd00"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.595250 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" (UID: "ce3ead5e-98d3-4719-b1d2-cda4d7b7b541"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.620498 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.620539 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.620551 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2k7l\" (UniqueName: \"kubernetes.io/projected/b36490b8-d5db-49c3-97e4-b734feaabd00-kube-api-access-h2k7l\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.620563 4946 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.620572 4946 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b36490b8-d5db-49c3-97e4-b734feaabd00-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.620580 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36490b8-d5db-49c3-97e4-b734feaabd00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.620589 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.621018 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.621067 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts podName:85a6dcc4-23cc-4f57-8714-06d0ef25c680 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:12.62105182 +0000 UTC m=+1565.417741929 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts") pod "barbicaneae6-account-delete-l9467" (UID: "85a6dcc4-23cc-4f57-8714-06d0ef25c680") : configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.698618 4946 scope.go:117] "RemoveContainer" containerID="6c046b12932defb2b1b37ab4cba5b712c6fb5e6d71e6a3ef814f1586ed8b402d" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.721664 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.728974 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.729042 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts podName:2f4c6455-fec3-41e2-bf0a-3595e34d63aa nodeName:}" failed. No retries permitted until 2025-12-03 07:16:12.729026932 +0000 UTC m=+1565.525717041 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts") pod "novaapi2e82-account-delete-77lgb" (UID: "2f4c6455-fec3-41e2-bf0a-3595e34d63aa") : configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.729356 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.729382 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts podName:0bc0555a-6c39-4e57-8c30-6a6fc422376a nodeName:}" failed. No retries permitted until 2025-12-03 07:16:12.729374191 +0000 UTC m=+1565.526064300 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts") pod "novacell0a5c7-account-delete-j2fhk" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a") : configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.729427 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.729443 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts podName:b6e24ccc-b386-4d6e-9c67-cc95e47cafab nodeName:}" failed. No retries permitted until 2025-12-03 07:16:12.729437293 +0000 UTC m=+1565.526127402 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts") pod "neutron6e16-account-delete-snvqr" (UID: "b6e24ccc-b386-4d6e-9c67-cc95e47cafab") : configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.729465 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: E1203 07:16:10.729484 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts podName:572a50eb-42e9-4897-9925-4073b5f6a35d nodeName:}" failed. No retries permitted until 2025-12-03 07:16:12.729479124 +0000 UTC m=+1565.526169233 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts") pod "placement0960-account-delete-pc5x5" (UID: "572a50eb-42e9-4897-9925-4073b5f6a35d") : configmap "openstack-scripts" not found Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.733830 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.799939 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hc87l" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" probeResult="failure" output="command timed out" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.800749 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.842024 4946 scope.go:117] "RemoveContainer" containerID="86788e20897ae74696026dd8776710b0bda123a3f5dc62f4eebb35926ddd9811" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.868506 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.876964 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.888912 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hc87l" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" probeResult="failure" output=< Dec 03 07:16:10 crc kubenswrapper[4946]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Dec 03 07:16:10 crc kubenswrapper[4946]: > Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.896177 4946 scope.go:117] "RemoveContainer" containerID="364174d07480a501e011d47109a2fdb307a389eee25a760b44b2337057f852ec" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.912679 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.926203 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5125-account-delete-lrpcc"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.931808 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder5125-account-delete-lrpcc"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.934588 4946 scope.go:117] "RemoveContainer" containerID="29a3b6fe73a28a60ca8a3174c922f9babad002d47efe3d41eecf7fee894fc739" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.940910 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-69454fc688-t54w7"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.955687 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-69454fc688-t54w7"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.957274 4946 scope.go:117] "RemoveContainer" containerID="434d3710c0181e612e1826bfbdad35d34bd77749c41c0c2a77720e391d7d862e" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.964876 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.981986 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.984031 4946 scope.go:117] "RemoveContainer" containerID="6c6cec4e34e14e36e892d11434e257757b6502d3fb6f96b9a2e74b1f7a78898a" Dec 03 07:16:10 crc kubenswrapper[4946]: I1203 07:16:10.999917 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.005105 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.015949 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.025150 4946 scope.go:117] "RemoveContainer" containerID="5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.025258 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.030192 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.037115 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.043925 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6786b69d4-clggf"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.049818 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6786b69d4-clggf"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.130855 4946 scope.go:117] "RemoveContainer" containerID="5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6" Dec 03 07:16:11 crc kubenswrapper[4946]: E1203 07:16:11.131302 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6\": container with ID starting with 5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6 not found: ID does not exist" containerID="5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.131340 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6"} err="failed to get container status \"5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6\": rpc error: code = NotFound desc = could not find container \"5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6\": container with ID starting with 5bd7cecb491fd2e13e7708826e30ff2513d1e9486d1fbc999830e9ec8b2179a6 not found: ID does not exist" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.206026 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b36490b8-d5db-49c3-97e4-b734feaabd00","Type":"ContainerDied","Data":"e9075da2f08d4d16898fc1dabf6593b7ad42915749b7c7fe8fea116b6a042271"} Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.206360 4946 scope.go:117] "RemoveContainer" containerID="40db7579ead1028473e650e6ae3b6bc768d742221d67f47bd5ce52721da60990" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.206048 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.212544 4946 generic.go:334] "Generic (PLEG): container finished" podID="85593cc4-b6bd-44e9-a908-58f99f0c8756" containerID="a625c10f25213752b4f3ffbe138270f09d8ca2e6105053ba6f35c9e4d430ce6e" exitCode=0 Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.212598 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"85593cc4-b6bd-44e9-a908-58f99f0c8756","Type":"ContainerDied","Data":"a625c10f25213752b4f3ffbe138270f09d8ca2e6105053ba6f35c9e4d430ce6e"} Dec 03 07:16:11 crc kubenswrapper[4946]: E1203 07:16:11.250014 4946 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 03 07:16:11 crc kubenswrapper[4946]: E1203 07:16:11.250090 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data podName:854575dc-d77d-43fe-a7ff-495ef76ab456 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:19.250073494 +0000 UTC m=+1572.046763603 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data") pod "rabbitmq-server-0" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456") : configmap "rabbitmq-config-data" not found Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.271566 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.279513 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.437054 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.556624 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-combined-ca-bundle\") pod \"85593cc4-b6bd-44e9-a908-58f99f0c8756\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.556707 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-operator-scripts\") pod \"85593cc4-b6bd-44e9-a908-58f99f0c8756\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.556727 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-kolla-config\") pod \"85593cc4-b6bd-44e9-a908-58f99f0c8756\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.556795 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"85593cc4-b6bd-44e9-a908-58f99f0c8756\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.556835 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdcrr\" (UniqueName: \"kubernetes.io/projected/85593cc4-b6bd-44e9-a908-58f99f0c8756-kube-api-access-zdcrr\") pod \"85593cc4-b6bd-44e9-a908-58f99f0c8756\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.556880 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-default\") pod \"85593cc4-b6bd-44e9-a908-58f99f0c8756\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.556928 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-generated\") pod \"85593cc4-b6bd-44e9-a908-58f99f0c8756\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.556947 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-galera-tls-certs\") pod \"85593cc4-b6bd-44e9-a908-58f99f0c8756\" (UID: \"85593cc4-b6bd-44e9-a908-58f99f0c8756\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.557807 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "85593cc4-b6bd-44e9-a908-58f99f0c8756" (UID: "85593cc4-b6bd-44e9-a908-58f99f0c8756"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.558647 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85593cc4-b6bd-44e9-a908-58f99f0c8756" (UID: "85593cc4-b6bd-44e9-a908-58f99f0c8756"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.561357 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "85593cc4-b6bd-44e9-a908-58f99f0c8756" (UID: "85593cc4-b6bd-44e9-a908-58f99f0c8756"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.561533 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "85593cc4-b6bd-44e9-a908-58f99f0c8756" (UID: "85593cc4-b6bd-44e9-a908-58f99f0c8756"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.564601 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85593cc4-b6bd-44e9-a908-58f99f0c8756-kube-api-access-zdcrr" (OuterVolumeSpecName: "kube-api-access-zdcrr") pod "85593cc4-b6bd-44e9-a908-58f99f0c8756" (UID: "85593cc4-b6bd-44e9-a908-58f99f0c8756"). InnerVolumeSpecName "kube-api-access-zdcrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.580466 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "85593cc4-b6bd-44e9-a908-58f99f0c8756" (UID: "85593cc4-b6bd-44e9-a908-58f99f0c8756"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.591118 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85593cc4-b6bd-44e9-a908-58f99f0c8756" (UID: "85593cc4-b6bd-44e9-a908-58f99f0c8756"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.606692 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e0b4114-0121-48cc-9439-bc5d350aeb44" path="/var/lib/kubelet/pods/0e0b4114-0121-48cc-9439-bc5d350aeb44/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.607176 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e57ed09-f7ce-403e-9cb5-676ae1c64117" path="/var/lib/kubelet/pods/0e57ed09-f7ce-403e-9cb5-676ae1c64117/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.607623 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" path="/var/lib/kubelet/pods/10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.608966 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" path="/var/lib/kubelet/pods/192c2964-d6c9-404e-8cca-cac01c750c5b/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.609550 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" path="/var/lib/kubelet/pods/4dc4eff7-f011-4265-b9dd-4e6bcdcb08df/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.610104 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="761f2199-bf20-487a-a18d-172c65a4b515" path="/var/lib/kubelet/pods/761f2199-bf20-487a-a18d-172c65a4b515/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.611610 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" path="/var/lib/kubelet/pods/7758850c-e98d-4bb6-bd84-31584a9bcdb6/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.612566 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "85593cc4-b6bd-44e9-a908-58f99f0c8756" (UID: "85593cc4-b6bd-44e9-a908-58f99f0c8756"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.612715 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" path="/var/lib/kubelet/pods/8a3521a3-0f30-4afe-b5f8-efeb6f838e72/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.614532 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b36490b8-d5db-49c3-97e4-b734feaabd00" path="/var/lib/kubelet/pods/b36490b8-d5db-49c3-97e4-b734feaabd00/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.615103 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baf286a1-0b76-4e66-817c-7d7ce84edaa2" path="/var/lib/kubelet/pods/baf286a1-0b76-4e66-817c-7d7ce84edaa2/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.618540 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" path="/var/lib/kubelet/pods/bd1d9c96-a0f4-43db-a686-2b320c1948ae/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.619431 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" path="/var/lib/kubelet/pods/ce3ead5e-98d3-4719-b1d2-cda4d7b7b541/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.620490 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5415d2e-cf73-4dd8-900d-211b957e829f" path="/var/lib/kubelet/pods/e5415d2e-cf73-4dd8-900d-211b957e829f/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.621131 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f69ee496-d325-40e0-81ff-0fb3625a775a" path="/var/lib/kubelet/pods/f69ee496-d325-40e0-81ff-0fb3625a775a/volumes" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.658749 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.658776 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdcrr\" (UniqueName: \"kubernetes.io/projected/85593cc4-b6bd-44e9-a908-58f99f0c8756-kube-api-access-zdcrr\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.658786 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.658794 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/85593cc4-b6bd-44e9-a908-58f99f0c8756-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.658803 4946 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.658812 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85593cc4-b6bd-44e9-a908-58f99f0c8756-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.658820 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.658827 4946 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85593cc4-b6bd-44e9-a908-58f99f0c8756-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.679025 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.740877 4946 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="04384280-d303-475d-8abc-c9c957db7fd9" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.195:6080/vnc_lite.html\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.759881 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.931929 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_80557a83-debb-4e3a-a2d0-bb5ac72d824c/ovn-northd/0.log" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.932049 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.962397 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-config\") pod \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.962515 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-combined-ca-bundle\") pod \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.962556 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-scripts\") pod \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.962598 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-rundir\") pod \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.962657 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-northd-tls-certs\") pod \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.962735 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-metrics-certs-tls-certs\") pod \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.962816 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrpfz\" (UniqueName: \"kubernetes.io/projected/80557a83-debb-4e3a-a2d0-bb5ac72d824c-kube-api-access-mrpfz\") pod \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.963081 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-config" (OuterVolumeSpecName: "config") pod "80557a83-debb-4e3a-a2d0-bb5ac72d824c" (UID: "80557a83-debb-4e3a-a2d0-bb5ac72d824c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.963161 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-scripts" (OuterVolumeSpecName: "scripts") pod "80557a83-debb-4e3a-a2d0-bb5ac72d824c" (UID: "80557a83-debb-4e3a-a2d0-bb5ac72d824c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.963486 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.963504 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80557a83-debb-4e3a-a2d0-bb5ac72d824c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.964329 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "80557a83-debb-4e3a-a2d0-bb5ac72d824c" (UID: "80557a83-debb-4e3a-a2d0-bb5ac72d824c"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.983587 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80557a83-debb-4e3a-a2d0-bb5ac72d824c-kube-api-access-mrpfz" (OuterVolumeSpecName: "kube-api-access-mrpfz") pod "80557a83-debb-4e3a-a2d0-bb5ac72d824c" (UID: "80557a83-debb-4e3a-a2d0-bb5ac72d824c"). InnerVolumeSpecName "kube-api-access-mrpfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.985047 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 07:16:11 crc kubenswrapper[4946]: I1203 07:16:11.998901 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80557a83-debb-4e3a-a2d0-bb5ac72d824c" (UID: "80557a83-debb-4e3a-a2d0-bb5ac72d824c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.053787 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "80557a83-debb-4e3a-a2d0-bb5ac72d824c" (UID: "80557a83-debb-4e3a-a2d0-bb5ac72d824c"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064407 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqpbx\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-kube-api-access-rqpbx\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064415 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "80557a83-debb-4e3a-a2d0-bb5ac72d824c" (UID: "80557a83-debb-4e3a-a2d0-bb5ac72d824c"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064440 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-tls\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064532 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064614 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-metrics-certs-tls-certs\") pod \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\" (UID: \"80557a83-debb-4e3a-a2d0-bb5ac72d824c\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064650 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-confd\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064689 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064797 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-erlang-cookie\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: W1203 07:16:12.064815 4946 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/80557a83-debb-4e3a-a2d0-bb5ac72d824c/volumes/kubernetes.io~secret/metrics-certs-tls-certs Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064830 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-plugins\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064838 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "80557a83-debb-4e3a-a2d0-bb5ac72d824c" (UID: "80557a83-debb-4e3a-a2d0-bb5ac72d824c"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064941 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/854575dc-d77d-43fe-a7ff-495ef76ab456-erlang-cookie-secret\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.064975 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-plugins-conf\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.065003 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-server-conf\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.065059 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/854575dc-d77d-43fe-a7ff-495ef76ab456-pod-info\") pod \"854575dc-d77d-43fe-a7ff-495ef76ab456\" (UID: \"854575dc-d77d-43fe-a7ff-495ef76ab456\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.066506 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.067077 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.068136 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.068483 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrpfz\" (UniqueName: \"kubernetes.io/projected/80557a83-debb-4e3a-a2d0-bb5ac72d824c-kube-api-access-mrpfz\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.068518 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.068548 4946 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.068562 4946 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.068574 4946 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/80557a83-debb-4e3a-a2d0-bb5ac72d824c-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.068488 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.068554 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-kube-api-access-rqpbx" (OuterVolumeSpecName: "kube-api-access-rqpbx") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "kube-api-access-rqpbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.069206 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.071336 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/854575dc-d77d-43fe-a7ff-495ef76ab456-pod-info" (OuterVolumeSpecName: "pod-info") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.072483 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/854575dc-d77d-43fe-a7ff-495ef76ab456-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.105903 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data" (OuterVolumeSpecName: "config-data") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.109335 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-server-conf" (OuterVolumeSpecName: "server-conf") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.170895 4946 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/854575dc-d77d-43fe-a7ff-495ef76ab456-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.170933 4946 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.170946 4946 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.170959 4946 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/854575dc-d77d-43fe-a7ff-495ef76ab456-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.170971 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqpbx\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-kube-api-access-rqpbx\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.171010 4946 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.171045 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.171060 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/854575dc-d77d-43fe-a7ff-495ef76ab456-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.171073 4946 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.171084 4946 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.192548 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "854575dc-d77d-43fe-a7ff-495ef76ab456" (UID: "854575dc-d77d-43fe-a7ff-495ef76ab456"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.194224 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.213746 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.241597 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_80557a83-debb-4e3a-a2d0-bb5ac72d824c/ovn-northd/0.log" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.241641 4946 generic.go:334] "Generic (PLEG): container finished" podID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerID="1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d" exitCode=139 Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.241686 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"80557a83-debb-4e3a-a2d0-bb5ac72d824c","Type":"ContainerDied","Data":"1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d"} Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.241712 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"80557a83-debb-4e3a-a2d0-bb5ac72d824c","Type":"ContainerDied","Data":"5ef0b652acc68849990842cb7fa135cf01021eef075d817329b9cb84a532cf80"} Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.241761 4946 scope.go:117] "RemoveContainer" containerID="6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.241858 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.247776 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"85593cc4-b6bd-44e9-a908-58f99f0c8756","Type":"ContainerDied","Data":"465c4b27ada46636f29c78d4626fd8ed46110914b69a336eb96d1ee7a96f4b39"} Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.247885 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.249358 4946 generic.go:334] "Generic (PLEG): container finished" podID="e30f3ad7-043b-4212-a496-656c3447b509" containerID="6f451caacded099e931e441c82a111b08382c08a1a8b89ed90e33967d206ea41" exitCode=0 Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.249406 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b848b8657-gfw2m" event={"ID":"e30f3ad7-043b-4212-a496-656c3447b509","Type":"ContainerDied","Data":"6f451caacded099e931e441c82a111b08382c08a1a8b89ed90e33967d206ea41"} Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.251394 4946 generic.go:334] "Generic (PLEG): container finished" podID="2f821218-d2a4-4f97-a924-199c056c7c55" containerID="7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e" exitCode=0 Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.251440 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2f821218-d2a4-4f97-a924-199c056c7c55","Type":"ContainerDied","Data":"7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e"} Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.251462 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2f821218-d2a4-4f97-a924-199c056c7c55","Type":"ContainerDied","Data":"85a4854921c2c14df675d2f527a4f30ec3f96f81ee2ccb5e5c48d858434366c3"} Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.251513 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.253646 4946 generic.go:334] "Generic (PLEG): container finished" podID="854575dc-d77d-43fe-a7ff-495ef76ab456" containerID="2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32" exitCode=0 Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.253670 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"854575dc-d77d-43fe-a7ff-495ef76ab456","Type":"ContainerDied","Data":"2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32"} Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.253685 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"854575dc-d77d-43fe-a7ff-495ef76ab456","Type":"ContainerDied","Data":"14aafb403e6899be5493858d2c6fb52724774ca8c09d0f1db1e4099cffb954ff"} Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.253694 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.272516 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.272620 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-plugins\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.272730 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-erlang-cookie\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.274300 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f821218-d2a4-4f97-a924-199c056c7c55-erlang-cookie-secret\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.274334 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.274374 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-tls\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.274400 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f821218-d2a4-4f97-a924-199c056c7c55-pod-info\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.274440 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-plugins-conf\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.274477 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-confd\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.274503 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5jf6\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-kube-api-access-h5jf6\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.274545 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-server-conf\") pod \"2f821218-d2a4-4f97-a924-199c056c7c55\" (UID: \"2f821218-d2a4-4f97-a924-199c056c7c55\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.275030 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.275050 4946 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/854575dc-d77d-43fe-a7ff-495ef76ab456-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.290788 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.302663 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.303346 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.303708 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.306133 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.306195 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.311808 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.317318 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.318224 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.322046 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-kube-api-access-h5jf6" (OuterVolumeSpecName: "kube-api-access-h5jf6") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "kube-api-access-h5jf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.322853 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f821218-d2a4-4f97-a924-199c056c7c55-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.323319 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2f821218-d2a4-4f97-a924-199c056c7c55-pod-info" (OuterVolumeSpecName: "pod-info") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.324536 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.335142 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data" (OuterVolumeSpecName: "config-data") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.337014 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.339682 4946 scope.go:117] "RemoveContainer" containerID="1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.365730 4946 scope.go:117] "RemoveContainer" containerID="6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82" Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.366355 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82\": container with ID starting with 6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82 not found: ID does not exist" containerID="6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.366387 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82"} err="failed to get container status \"6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82\": rpc error: code = NotFound desc = could not find container \"6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82\": container with ID starting with 6c80f5123c7486012ce4d6091fdbe00762ca9598eb315e0f36cebc56d75c8e82 not found: ID does not exist" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.366407 4946 scope.go:117] "RemoveContainer" containerID="1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.366815 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-server-conf" (OuterVolumeSpecName: "server-conf") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.366925 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d\": container with ID starting with 1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d not found: ID does not exist" containerID="1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.366945 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d"} err="failed to get container status \"1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d\": rpc error: code = NotFound desc = could not find container \"1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d\": container with ID starting with 1246c6762a3565391cddc36261e027f937737bcae00909e8b6619827feaf902d not found: ID does not exist" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.366959 4946 scope.go:117] "RemoveContainer" containerID="a625c10f25213752b4f3ffbe138270f09d8ca2e6105053ba6f35c9e4d430ce6e" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376494 4946 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f821218-d2a4-4f97-a924-199c056c7c55-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376516 4946 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376525 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5jf6\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-kube-api-access-h5jf6\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376535 4946 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376559 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376568 4946 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376576 4946 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376587 4946 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f821218-d2a4-4f97-a924-199c056c7c55-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376596 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f821218-d2a4-4f97-a924-199c056c7c55-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.376604 4946 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.386418 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2f821218-d2a4-4f97-a924-199c056c7c55" (UID: "2f821218-d2a4-4f97-a924-199c056c7c55"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.392472 4946 scope.go:117] "RemoveContainer" containerID="2415dc5d63b46387779f532cd38fe4a49aee041aa819437a99561449e7e5821c" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.419284 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.426662 4946 scope.go:117] "RemoveContainer" containerID="7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.446707 4946 scope.go:117] "RemoveContainer" containerID="15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.479110 4946 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f821218-d2a4-4f97-a924-199c056c7c55-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.479144 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.481595 4946 scope.go:117] "RemoveContainer" containerID="7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e" Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.482053 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e\": container with ID starting with 7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e not found: ID does not exist" containerID="7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.482128 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e"} err="failed to get container status \"7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e\": rpc error: code = NotFound desc = could not find container \"7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e\": container with ID starting with 7ad1c9b5e6edf58ec0edef2742496688ed96b12cefeaec303742a152aad4053e not found: ID does not exist" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.482149 4946 scope.go:117] "RemoveContainer" containerID="15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d" Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.482577 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d\": container with ID starting with 15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d not found: ID does not exist" containerID="15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.482598 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d"} err="failed to get container status \"15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d\": rpc error: code = NotFound desc = could not find container \"15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d\": container with ID starting with 15e85a8b8feb805a4612d477af412ab81ff994df0f5b1177c556797447e34f3d not found: ID does not exist" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.482611 4946 scope.go:117] "RemoveContainer" containerID="2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.499580 4946 scope.go:117] "RemoveContainer" containerID="5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.517705 4946 scope.go:117] "RemoveContainer" containerID="2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32" Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.518216 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32\": container with ID starting with 2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32 not found: ID does not exist" containerID="2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.518253 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32"} err="failed to get container status \"2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32\": rpc error: code = NotFound desc = could not find container \"2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32\": container with ID starting with 2f645c095cf6e37c9fb6d0543f0585954de3e1489bb253835fd974e1e72a1e32 not found: ID does not exist" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.518280 4946 scope.go:117] "RemoveContainer" containerID="5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78" Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.518664 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78\": container with ID starting with 5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78 not found: ID does not exist" containerID="5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.518689 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78"} err="failed to get container status \"5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78\": rpc error: code = NotFound desc = could not find container \"5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78\": container with ID starting with 5f4693ac29b55bf22e47d3001ab3d37016c3950b50795dfddbd77df3d380dd78 not found: ID does not exist" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.543776 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.579600 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw6qg\" (UniqueName: \"kubernetes.io/projected/e30f3ad7-043b-4212-a496-656c3447b509-kube-api-access-bw6qg\") pod \"e30f3ad7-043b-4212-a496-656c3447b509\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.579656 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-internal-tls-certs\") pod \"e30f3ad7-043b-4212-a496-656c3447b509\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.579821 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-config-data\") pod \"e30f3ad7-043b-4212-a496-656c3447b509\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.580019 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-public-tls-certs\") pod \"e30f3ad7-043b-4212-a496-656c3447b509\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.580039 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-scripts\") pod \"e30f3ad7-043b-4212-a496-656c3447b509\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.580061 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-combined-ca-bundle\") pod \"e30f3ad7-043b-4212-a496-656c3447b509\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.580132 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-credential-keys\") pod \"e30f3ad7-043b-4212-a496-656c3447b509\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.580151 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-fernet-keys\") pod \"e30f3ad7-043b-4212-a496-656c3447b509\" (UID: \"e30f3ad7-043b-4212-a496-656c3447b509\") " Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.594276 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e30f3ad7-043b-4212-a496-656c3447b509-kube-api-access-bw6qg" (OuterVolumeSpecName: "kube-api-access-bw6qg") pod "e30f3ad7-043b-4212-a496-656c3447b509" (UID: "e30f3ad7-043b-4212-a496-656c3447b509"). InnerVolumeSpecName "kube-api-access-bw6qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.605991 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.612493 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.613400 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e30f3ad7-043b-4212-a496-656c3447b509" (UID: "e30f3ad7-043b-4212-a496-656c3447b509"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.626694 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-scripts" (OuterVolumeSpecName: "scripts") pod "e30f3ad7-043b-4212-a496-656c3447b509" (UID: "e30f3ad7-043b-4212-a496-656c3447b509"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.645269 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e30f3ad7-043b-4212-a496-656c3447b509" (UID: "e30f3ad7-043b-4212-a496-656c3447b509"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.670082 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e30f3ad7-043b-4212-a496-656c3447b509" (UID: "e30f3ad7-043b-4212-a496-656c3447b509"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.675447 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e30f3ad7-043b-4212-a496-656c3447b509" (UID: "e30f3ad7-043b-4212-a496-656c3447b509"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.680706 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-config-data" (OuterVolumeSpecName: "config-data") pod "e30f3ad7-043b-4212-a496-656c3447b509" (UID: "e30f3ad7-043b-4212-a496-656c3447b509"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.681812 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.681897 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.681948 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.682018 4946 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.682068 4946 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.682138 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw6qg\" (UniqueName: \"kubernetes.io/projected/e30f3ad7-043b-4212-a496-656c3447b509-kube-api-access-bw6qg\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.682189 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.682974 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.683076 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts podName:85a6dcc4-23cc-4f57-8714-06d0ef25c680 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:16.683060795 +0000 UTC m=+1569.479750904 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts") pod "barbicaneae6-account-delete-l9467" (UID: "85a6dcc4-23cc-4f57-8714-06d0ef25c680") : configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.692692 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e30f3ad7-043b-4212-a496-656c3447b509" (UID: "e30f3ad7-043b-4212-a496-656c3447b509"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:12 crc kubenswrapper[4946]: I1203 07:16:12.783533 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e30f3ad7-043b-4212-a496-656c3447b509-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.783615 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.783660 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts podName:0bc0555a-6c39-4e57-8c30-6a6fc422376a nodeName:}" failed. No retries permitted until 2025-12-03 07:16:16.783646057 +0000 UTC m=+1569.580336166 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts") pod "novacell0a5c7-account-delete-j2fhk" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a") : configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.784315 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.784344 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts podName:b6e24ccc-b386-4d6e-9c67-cc95e47cafab nodeName:}" failed. No retries permitted until 2025-12-03 07:16:16.784336616 +0000 UTC m=+1569.581026725 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts") pod "neutron6e16-account-delete-snvqr" (UID: "b6e24ccc-b386-4d6e-9c67-cc95e47cafab") : configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.784368 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.784385 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts podName:572a50eb-42e9-4897-9925-4073b5f6a35d nodeName:}" failed. No retries permitted until 2025-12-03 07:16:16.784380737 +0000 UTC m=+1569.581070846 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts") pod "placement0960-account-delete-pc5x5" (UID: "572a50eb-42e9-4897-9925-4073b5f6a35d") : configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.784407 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:12 crc kubenswrapper[4946]: E1203 07:16:12.784422 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts podName:2f4c6455-fec3-41e2-bf0a-3595e34d63aa nodeName:}" failed. No retries permitted until 2025-12-03 07:16:16.784417728 +0000 UTC m=+1569.581107837 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts") pod "novaapi2e82-account-delete-77lgb" (UID: "2f4c6455-fec3-41e2-bf0a-3595e34d63aa") : configmap "openstack-scripts" not found Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.099890 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:16:13 crc kubenswrapper[4946]: E1203 07:16:13.114352 4946 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 03 07:16:13 crc kubenswrapper[4946]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-03T07:16:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 03 07:16:13 crc kubenswrapper[4946]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 03 07:16:13 crc kubenswrapper[4946]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-hc87l" message=< Dec 03 07:16:13 crc kubenswrapper[4946]: Exiting ovn-controller (1) [FAILED] Dec 03 07:16:13 crc kubenswrapper[4946]: Killing ovn-controller (1) [ OK ] Dec 03 07:16:13 crc kubenswrapper[4946]: Killing ovn-controller (1) with SIGKILL [ OK ] Dec 03 07:16:13 crc kubenswrapper[4946]: 2025-12-03T07:16:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 03 07:16:13 crc kubenswrapper[4946]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 03 07:16:13 crc kubenswrapper[4946]: > Dec 03 07:16:13 crc kubenswrapper[4946]: E1203 07:16:13.114420 4946 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 03 07:16:13 crc kubenswrapper[4946]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-03T07:16:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 03 07:16:13 crc kubenswrapper[4946]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 03 07:16:13 crc kubenswrapper[4946]: > pod="openstack/ovn-controller-hc87l" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" containerID="cri-o://8293f410bee38eef4e13c919564b09535995b3ce313122377c14666333bd220b" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.114457 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-hc87l" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" containerID="cri-o://8293f410bee38eef4e13c919564b09535995b3ce313122377c14666333bd220b" gracePeriod=22 Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.194826 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.200672 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data-custom\") pod \"26ac177b-b889-43e7-888e-8ed02fb3ac72\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.200753 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzd5s\" (UniqueName: \"kubernetes.io/projected/26ac177b-b889-43e7-888e-8ed02fb3ac72-kube-api-access-kzd5s\") pod \"26ac177b-b889-43e7-888e-8ed02fb3ac72\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.200779 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data\") pod \"26ac177b-b889-43e7-888e-8ed02fb3ac72\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.200805 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ac177b-b889-43e7-888e-8ed02fb3ac72-logs\") pod \"26ac177b-b889-43e7-888e-8ed02fb3ac72\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.200837 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-combined-ca-bundle\") pod \"26ac177b-b889-43e7-888e-8ed02fb3ac72\" (UID: \"26ac177b-b889-43e7-888e-8ed02fb3ac72\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.201965 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26ac177b-b889-43e7-888e-8ed02fb3ac72-logs" (OuterVolumeSpecName: "logs") pod "26ac177b-b889-43e7-888e-8ed02fb3ac72" (UID: "26ac177b-b889-43e7-888e-8ed02fb3ac72"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.214131 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26ac177b-b889-43e7-888e-8ed02fb3ac72-kube-api-access-kzd5s" (OuterVolumeSpecName: "kube-api-access-kzd5s") pod "26ac177b-b889-43e7-888e-8ed02fb3ac72" (UID: "26ac177b-b889-43e7-888e-8ed02fb3ac72"). InnerVolumeSpecName "kube-api-access-kzd5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.215217 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "26ac177b-b889-43e7-888e-8ed02fb3ac72" (UID: "26ac177b-b889-43e7-888e-8ed02fb3ac72"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.243165 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26ac177b-b889-43e7-888e-8ed02fb3ac72" (UID: "26ac177b-b889-43e7-888e-8ed02fb3ac72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.272134 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-hc87l_be4f07e5-c606-4eb4-95cc-5e5c7dcb195d/ovn-controller/0.log" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.272181 4946 generic.go:334] "Generic (PLEG): container finished" podID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerID="8293f410bee38eef4e13c919564b09535995b3ce313122377c14666333bd220b" exitCode=137 Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.272238 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hc87l" event={"ID":"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d","Type":"ContainerDied","Data":"8293f410bee38eef4e13c919564b09535995b3ce313122377c14666333bd220b"} Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.282254 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.283687 4946 generic.go:334] "Generic (PLEG): container finished" podID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerID="ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f" exitCode=0 Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.283845 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-559fcd84cc-lhklt" event={"ID":"26ac177b-b889-43e7-888e-8ed02fb3ac72","Type":"ContainerDied","Data":"ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f"} Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.283885 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-559fcd84cc-lhklt" event={"ID":"26ac177b-b889-43e7-888e-8ed02fb3ac72","Type":"ContainerDied","Data":"50cd82febee37545cca600c5780372aa7dfb6ae8d5ff07654a79fd9977964e38"} Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.283903 4946 scope.go:117] "RemoveContainer" containerID="ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.283906 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-559fcd84cc-lhklt" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.294986 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data" (OuterVolumeSpecName: "config-data") pod "26ac177b-b889-43e7-888e-8ed02fb3ac72" (UID: "26ac177b-b889-43e7-888e-8ed02fb3ac72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.297568 4946 generic.go:334] "Generic (PLEG): container finished" podID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerID="733bbc32986c6be4c2390635eddd73d705ca1fffc3e0db20cc69f0e5434d3516" exitCode=0 Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.297653 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d9d0798c-6a8a-4df3-9270-c75d49a2379c","Type":"ContainerDied","Data":"733bbc32986c6be4c2390635eddd73d705ca1fffc3e0db20cc69f0e5434d3516"} Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.297803 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.303840 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data\") pod \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.303961 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-logs\") pod \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.303998 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data-custom\") pod \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.304023 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2g8l\" (UniqueName: \"kubernetes.io/projected/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-kube-api-access-h2g8l\") pod \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.304047 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-combined-ca-bundle\") pod \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\" (UID: \"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.304457 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-logs" (OuterVolumeSpecName: "logs") pod "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" (UID: "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.304818 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.304843 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.304860 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzd5s\" (UniqueName: \"kubernetes.io/projected/26ac177b-b889-43e7-888e-8ed02fb3ac72-kube-api-access-kzd5s\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.304878 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.304960 4946 generic.go:334] "Generic (PLEG): container finished" podID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerID="4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f" exitCode=0 Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.305031 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" event={"ID":"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e","Type":"ContainerDied","Data":"4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f"} Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.305086 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.305088 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f55c455cd-jdhgw" event={"ID":"92ce5650-798f-4d12-bbd2-4ad5f2c47c5e","Type":"ContainerDied","Data":"068f56e9cdc616e48e6fcc81467f0f7484fea27c9e42d09a6874ef75ea9bbcd8"} Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.305295 4946 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26ac177b-b889-43e7-888e-8ed02fb3ac72-logs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.305426 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ac177b-b889-43e7-888e-8ed02fb3ac72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.306955 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b848b8657-gfw2m" event={"ID":"e30f3ad7-043b-4212-a496-656c3447b509","Type":"ContainerDied","Data":"f14ba74c11fbec5585472424d5c2360f65091f851ecf574c28a4a6adc40dd3aa"} Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.307023 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b848b8657-gfw2m" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.307177 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" (UID: "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.313719 4946 scope.go:117] "RemoveContainer" containerID="27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.315157 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-kube-api-access-h2g8l" (OuterVolumeSpecName: "kube-api-access-h2g8l") pod "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" (UID: "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e"). InnerVolumeSpecName "kube-api-access-h2g8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.352923 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" (UID: "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.357129 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6b848b8657-gfw2m"] Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.357247 4946 scope.go:117] "RemoveContainer" containerID="ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f" Dec 03 07:16:13 crc kubenswrapper[4946]: E1203 07:16:13.358846 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f\": container with ID starting with ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f not found: ID does not exist" containerID="ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.358878 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f"} err="failed to get container status \"ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f\": rpc error: code = NotFound desc = could not find container \"ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f\": container with ID starting with ca2c0ff415ea817b11cad032eedf0ff419f7ba5e423822e91087c4768379928f not found: ID does not exist" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.358899 4946 scope.go:117] "RemoveContainer" containerID="27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060" Dec 03 07:16:13 crc kubenswrapper[4946]: E1203 07:16:13.360435 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060\": container with ID starting with 27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060 not found: ID does not exist" containerID="27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.360457 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060"} err="failed to get container status \"27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060\": rpc error: code = NotFound desc = could not find container \"27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060\": container with ID starting with 27bde9320f1e536118e3ea073e636912385f952bb697582f3c643fc3abc41060 not found: ID does not exist" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.360471 4946 scope.go:117] "RemoveContainer" containerID="b88d4d8819b799ee0958eb33b8f4190198f52e90c0717a81284a3e8f801ddb08" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.370040 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6b848b8657-gfw2m"] Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.378702 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data" (OuterVolumeSpecName: "config-data") pod "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" (UID: "92ce5650-798f-4d12-bbd2-4ad5f2c47c5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.393572 4946 scope.go:117] "RemoveContainer" containerID="9242e69cc08af9bee9b0d9c36b79f79805314218c4453cc7949a223cbc6ad191" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.405865 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-sg-core-conf-yaml\") pod \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.405962 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-ceilometer-tls-certs\") pod \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.405988 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-run-httpd\") pod \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406017 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-combined-ca-bundle\") pod \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406044 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-log-httpd\") pod \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406089 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-config-data\") pod \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406143 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvn6t\" (UniqueName: \"kubernetes.io/projected/d9d0798c-6a8a-4df3-9270-c75d49a2379c-kube-api-access-vvn6t\") pod \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406171 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-scripts\") pod \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\" (UID: \"d9d0798c-6a8a-4df3-9270-c75d49a2379c\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406478 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406502 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d9d0798c-6a8a-4df3-9270-c75d49a2379c" (UID: "d9d0798c-6a8a-4df3-9270-c75d49a2379c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406516 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406566 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2g8l\" (UniqueName: \"kubernetes.io/projected/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-kube-api-access-h2g8l\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.406585 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.409589 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-scripts" (OuterVolumeSpecName: "scripts") pod "d9d0798c-6a8a-4df3-9270-c75d49a2379c" (UID: "d9d0798c-6a8a-4df3-9270-c75d49a2379c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.412202 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d0798c-6a8a-4df3-9270-c75d49a2379c-kube-api-access-vvn6t" (OuterVolumeSpecName: "kube-api-access-vvn6t") pod "d9d0798c-6a8a-4df3-9270-c75d49a2379c" (UID: "d9d0798c-6a8a-4df3-9270-c75d49a2379c"). InnerVolumeSpecName "kube-api-access-vvn6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.412482 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d9d0798c-6a8a-4df3-9270-c75d49a2379c" (UID: "d9d0798c-6a8a-4df3-9270-c75d49a2379c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.413284 4946 scope.go:117] "RemoveContainer" containerID="733bbc32986c6be4c2390635eddd73d705ca1fffc3e0db20cc69f0e5434d3516" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.425913 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d9d0798c-6a8a-4df3-9270-c75d49a2379c" (UID: "d9d0798c-6a8a-4df3-9270-c75d49a2379c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.434240 4946 scope.go:117] "RemoveContainer" containerID="fd75ea492b2e5d3b5ac427900e214c87df97693bea489cdea1450e4ac5fce395" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.443781 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d9d0798c-6a8a-4df3-9270-c75d49a2379c" (UID: "d9d0798c-6a8a-4df3-9270-c75d49a2379c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.457753 4946 scope.go:117] "RemoveContainer" containerID="4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f" Dec 03 07:16:13 crc kubenswrapper[4946]: E1203 07:16:13.467582 4946 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode30f3ad7_043b_4212_a496_656c3447b509.slice/crio-f14ba74c11fbec5585472424d5c2360f65091f851ecf574c28a4a6adc40dd3aa\": RecentStats: unable to find data in memory cache], [\"/system.slice/system-systemd\\\\x2dcoredump.slice/systemd-coredump@1-75247-0.service\": RecentStats: unable to find data in memory cache]" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.470637 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-hc87l_be4f07e5-c606-4eb4-95cc-5e5c7dcb195d/ovn-controller/0.log" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.470707 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hc87l" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.478666 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9d0798c-6a8a-4df3-9270-c75d49a2379c" (UID: "d9d0798c-6a8a-4df3-9270-c75d49a2379c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.486956 4946 scope.go:117] "RemoveContainer" containerID="444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.507454 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-scripts\") pod \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.507523 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-log-ovn\") pod \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.507562 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run\") pod \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.507672 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run-ovn\") pod \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.507733 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-combined-ca-bundle\") pod \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.507793 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-ovn-controller-tls-certs\") pod \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.508144 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-244bg\" (UniqueName: \"kubernetes.io/projected/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-kube-api-access-244bg\") pod \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\" (UID: \"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d\") " Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.508562 4946 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.508580 4946 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.508592 4946 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.508603 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.508613 4946 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d9d0798c-6a8a-4df3-9270-c75d49a2379c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.508624 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvn6t\" (UniqueName: \"kubernetes.io/projected/d9d0798c-6a8a-4df3-9270-c75d49a2379c-kube-api-access-vvn6t\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.508638 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.510131 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-config-data" (OuterVolumeSpecName: "config-data") pod "d9d0798c-6a8a-4df3-9270-c75d49a2379c" (UID: "d9d0798c-6a8a-4df3-9270-c75d49a2379c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.510201 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run" (OuterVolumeSpecName: "var-run") pod "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" (UID: "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.510506 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" (UID: "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.511645 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-scripts" (OuterVolumeSpecName: "scripts") pod "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" (UID: "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.511695 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" (UID: "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.514753 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-kube-api-access-244bg" (OuterVolumeSpecName: "kube-api-access-244bg") pod "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" (UID: "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d"). InnerVolumeSpecName "kube-api-access-244bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.521875 4946 scope.go:117] "RemoveContainer" containerID="4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f" Dec 03 07:16:13 crc kubenswrapper[4946]: E1203 07:16:13.525908 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f\": container with ID starting with 4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f not found: ID does not exist" containerID="4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.525960 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f"} err="failed to get container status \"4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f\": rpc error: code = NotFound desc = could not find container \"4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f\": container with ID starting with 4aa4e71ec3bbd7f7ce71ec54772eb05eb1aab1ba6f90789dc562cc810d83417f not found: ID does not exist" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.525994 4946 scope.go:117] "RemoveContainer" containerID="444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510" Dec 03 07:16:13 crc kubenswrapper[4946]: E1203 07:16:13.526439 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510\": container with ID starting with 444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510 not found: ID does not exist" containerID="444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.526506 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510"} err="failed to get container status \"444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510\": rpc error: code = NotFound desc = could not find container \"444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510\": container with ID starting with 444d5ef22ddcd8f74ef3f3d8835079dc45e9938daef1fbb9bc98f54b8895d510 not found: ID does not exist" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.526531 4946 scope.go:117] "RemoveContainer" containerID="6f451caacded099e931e441c82a111b08382c08a1a8b89ed90e33967d206ea41" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.535548 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" (UID: "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.568144 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" (UID: "be4f07e5-c606-4eb4-95cc-5e5c7dcb195d"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.609447 4946 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.609469 4946 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.609478 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9d0798c-6a8a-4df3-9270-c75d49a2379c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.609487 4946 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.609496 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.609505 4946 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.609514 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-244bg\" (UniqueName: \"kubernetes.io/projected/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-kube-api-access-244bg\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.609522 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.613919 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f821218-d2a4-4f97-a924-199c056c7c55" path="/var/lib/kubelet/pods/2f821218-d2a4-4f97-a924-199c056c7c55/volumes" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.614528 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" path="/var/lib/kubelet/pods/80557a83-debb-4e3a-a2d0-bb5ac72d824c/volumes" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.616002 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="854575dc-d77d-43fe-a7ff-495ef76ab456" path="/var/lib/kubelet/pods/854575dc-d77d-43fe-a7ff-495ef76ab456/volumes" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.616646 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85593cc4-b6bd-44e9-a908-58f99f0c8756" path="/var/lib/kubelet/pods/85593cc4-b6bd-44e9-a908-58f99f0c8756/volumes" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.617192 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e30f3ad7-043b-4212-a496-656c3447b509" path="/var/lib/kubelet/pods/e30f3ad7-043b-4212-a496-656c3447b509/volumes" Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.696881 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-559fcd84cc-lhklt"] Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.704325 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-559fcd84cc-lhklt"] Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.709242 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7f55c455cd-jdhgw"] Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.718915 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7f55c455cd-jdhgw"] Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.721005 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:16:13 crc kubenswrapper[4946]: I1203 07:16:13.727184 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 07:16:14 crc kubenswrapper[4946]: I1203 07:16:14.325594 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-hc87l_be4f07e5-c606-4eb4-95cc-5e5c7dcb195d/ovn-controller/0.log" Dec 03 07:16:14 crc kubenswrapper[4946]: I1203 07:16:14.325848 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hc87l" event={"ID":"be4f07e5-c606-4eb4-95cc-5e5c7dcb195d","Type":"ContainerDied","Data":"484cecebdb5b1d74f73d5421d9f200478f97e3f71bdc33a2c78f5251c475aada"} Dec 03 07:16:14 crc kubenswrapper[4946]: I1203 07:16:14.325882 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hc87l" Dec 03 07:16:14 crc kubenswrapper[4946]: I1203 07:16:14.325910 4946 scope.go:117] "RemoveContainer" containerID="8293f410bee38eef4e13c919564b09535995b3ce313122377c14666333bd220b" Dec 03 07:16:14 crc kubenswrapper[4946]: I1203 07:16:14.364146 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hc87l"] Dec 03 07:16:14 crc kubenswrapper[4946]: I1203 07:16:14.370117 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-hc87l"] Dec 03 07:16:15 crc kubenswrapper[4946]: E1203 07:16:15.187111 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:15 crc kubenswrapper[4946]: E1203 07:16:15.187499 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:15 crc kubenswrapper[4946]: E1203 07:16:15.188004 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:15 crc kubenswrapper[4946]: E1203 07:16:15.188060 4946 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:15 crc kubenswrapper[4946]: E1203 07:16:15.188927 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:15 crc kubenswrapper[4946]: E1203 07:16:15.194132 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:15 crc kubenswrapper[4946]: E1203 07:16:15.196026 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:15 crc kubenswrapper[4946]: E1203 07:16:15.196133 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:15 crc kubenswrapper[4946]: I1203 07:16:15.614181 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" path="/var/lib/kubelet/pods/26ac177b-b889-43e7-888e-8ed02fb3ac72/volumes" Dec 03 07:16:15 crc kubenswrapper[4946]: I1203 07:16:15.616688 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" path="/var/lib/kubelet/pods/92ce5650-798f-4d12-bbd2-4ad5f2c47c5e/volumes" Dec 03 07:16:15 crc kubenswrapper[4946]: I1203 07:16:15.618350 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" path="/var/lib/kubelet/pods/be4f07e5-c606-4eb4-95cc-5e5c7dcb195d/volumes" Dec 03 07:16:15 crc kubenswrapper[4946]: I1203 07:16:15.621238 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" path="/var/lib/kubelet/pods/d9d0798c-6a8a-4df3-9270-c75d49a2379c/volumes" Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.761421 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.761527 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts podName:85a6dcc4-23cc-4f57-8714-06d0ef25c680 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:24.761506019 +0000 UTC m=+1577.558196168 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts") pod "barbicaneae6-account-delete-l9467" (UID: "85a6dcc4-23cc-4f57-8714-06d0ef25c680") : configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.863193 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.863275 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts podName:0bc0555a-6c39-4e57-8c30-6a6fc422376a nodeName:}" failed. No retries permitted until 2025-12-03 07:16:24.863254723 +0000 UTC m=+1577.659944832 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts") pod "novacell0a5c7-account-delete-j2fhk" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a") : configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.863351 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.863524 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts podName:b6e24ccc-b386-4d6e-9c67-cc95e47cafab nodeName:}" failed. No retries permitted until 2025-12-03 07:16:24.863491579 +0000 UTC m=+1577.660181758 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts") pod "neutron6e16-account-delete-snvqr" (UID: "b6e24ccc-b386-4d6e-9c67-cc95e47cafab") : configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.863615 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.863795 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts podName:572a50eb-42e9-4897-9925-4073b5f6a35d nodeName:}" failed. No retries permitted until 2025-12-03 07:16:24.863728436 +0000 UTC m=+1577.660418585 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts") pod "placement0960-account-delete-pc5x5" (UID: "572a50eb-42e9-4897-9925-4073b5f6a35d") : configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.864315 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:16 crc kubenswrapper[4946]: E1203 07:16:16.864368 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts podName:2f4c6455-fec3-41e2-bf0a-3595e34d63aa nodeName:}" failed. No retries permitted until 2025-12-03 07:16:24.864357543 +0000 UTC m=+1577.661047652 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts") pod "novaapi2e82-account-delete-77lgb" (UID: "2f4c6455-fec3-41e2-bf0a-3595e34d63aa") : configmap "openstack-scripts" not found Dec 03 07:16:20 crc kubenswrapper[4946]: E1203 07:16:20.185404 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:20 crc kubenswrapper[4946]: E1203 07:16:20.187023 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:20 crc kubenswrapper[4946]: E1203 07:16:20.187028 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:20 crc kubenswrapper[4946]: E1203 07:16:20.189388 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:20 crc kubenswrapper[4946]: E1203 07:16:20.189379 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:20 crc kubenswrapper[4946]: E1203 07:16:20.189426 4946 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:20 crc kubenswrapper[4946]: E1203 07:16:20.191489 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:20 crc kubenswrapper[4946]: E1203 07:16:20.191555 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.039797 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.039909 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.039974 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.040924 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.041031 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" gracePeriod=600 Dec 03 07:16:23 crc kubenswrapper[4946]: E1203 07:16:23.194258 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.690786 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" exitCode=0 Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.690835 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113"} Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.690876 4946 scope.go:117] "RemoveContainer" containerID="a645c07ea34ddd66718ac1597141e393ef581a3ac2975f8c8d11ca9aa11d50ab" Dec 03 07:16:23 crc kubenswrapper[4946]: I1203 07:16:23.691430 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:16:23 crc kubenswrapper[4946]: E1203 07:16:23.691725 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.811415 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.811516 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts podName:85a6dcc4-23cc-4f57-8714-06d0ef25c680 nodeName:}" failed. No retries permitted until 2025-12-03 07:16:40.811490984 +0000 UTC m=+1593.608181163 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts") pod "barbicaneae6-account-delete-l9467" (UID: "85a6dcc4-23cc-4f57-8714-06d0ef25c680") : configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.912913 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.913048 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.913125 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.913153 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.913287 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts podName:2f4c6455-fec3-41e2-bf0a-3595e34d63aa nodeName:}" failed. No retries permitted until 2025-12-03 07:16:40.913270379 +0000 UTC m=+1593.709960478 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts") pod "novaapi2e82-account-delete-77lgb" (UID: "2f4c6455-fec3-41e2-bf0a-3595e34d63aa") : configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.913465 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts podName:b6e24ccc-b386-4d6e-9c67-cc95e47cafab nodeName:}" failed. No retries permitted until 2025-12-03 07:16:40.913446194 +0000 UTC m=+1593.710136323 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts") pod "neutron6e16-account-delete-snvqr" (UID: "b6e24ccc-b386-4d6e-9c67-cc95e47cafab") : configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.913494 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts podName:572a50eb-42e9-4897-9925-4073b5f6a35d nodeName:}" failed. No retries permitted until 2025-12-03 07:16:40.913480784 +0000 UTC m=+1593.710170913 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts") pod "placement0960-account-delete-pc5x5" (UID: "572a50eb-42e9-4897-9925-4073b5f6a35d") : configmap "openstack-scripts" not found Dec 03 07:16:24 crc kubenswrapper[4946]: E1203 07:16:24.913550 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts podName:0bc0555a-6c39-4e57-8c30-6a6fc422376a nodeName:}" failed. No retries permitted until 2025-12-03 07:16:40.913538766 +0000 UTC m=+1593.710228895 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts") pod "novacell0a5c7-account-delete-j2fhk" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a") : configmap "openstack-scripts" not found Dec 03 07:16:25 crc kubenswrapper[4946]: E1203 07:16:25.186265 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:25 crc kubenswrapper[4946]: E1203 07:16:25.186979 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:25 crc kubenswrapper[4946]: E1203 07:16:25.187467 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:25 crc kubenswrapper[4946]: E1203 07:16:25.187720 4946 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:25 crc kubenswrapper[4946]: E1203 07:16:25.187612 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:25 crc kubenswrapper[4946]: E1203 07:16:25.190091 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:25 crc kubenswrapper[4946]: E1203 07:16:25.194175 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:25 crc kubenswrapper[4946]: E1203 07:16:25.194615 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.348109 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.362865 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-internal-tls-certs\") pod \"cc385f93-2111-42cf-93be-d4cc82225570\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.362943 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-combined-ca-bundle\") pod \"cc385f93-2111-42cf-93be-d4cc82225570\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.363157 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-httpd-config\") pod \"cc385f93-2111-42cf-93be-d4cc82225570\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.363240 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sls7d\" (UniqueName: \"kubernetes.io/projected/cc385f93-2111-42cf-93be-d4cc82225570-kube-api-access-sls7d\") pod \"cc385f93-2111-42cf-93be-d4cc82225570\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.363307 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-config\") pod \"cc385f93-2111-42cf-93be-d4cc82225570\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.363373 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-ovndb-tls-certs\") pod \"cc385f93-2111-42cf-93be-d4cc82225570\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.363437 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-public-tls-certs\") pod \"cc385f93-2111-42cf-93be-d4cc82225570\" (UID: \"cc385f93-2111-42cf-93be-d4cc82225570\") " Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.373195 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "cc385f93-2111-42cf-93be-d4cc82225570" (UID: "cc385f93-2111-42cf-93be-d4cc82225570"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.388659 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc385f93-2111-42cf-93be-d4cc82225570-kube-api-access-sls7d" (OuterVolumeSpecName: "kube-api-access-sls7d") pod "cc385f93-2111-42cf-93be-d4cc82225570" (UID: "cc385f93-2111-42cf-93be-d4cc82225570"). InnerVolumeSpecName "kube-api-access-sls7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.428590 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cc385f93-2111-42cf-93be-d4cc82225570" (UID: "cc385f93-2111-42cf-93be-d4cc82225570"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.435361 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-config" (OuterVolumeSpecName: "config") pod "cc385f93-2111-42cf-93be-d4cc82225570" (UID: "cc385f93-2111-42cf-93be-d4cc82225570"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.441371 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc385f93-2111-42cf-93be-d4cc82225570" (UID: "cc385f93-2111-42cf-93be-d4cc82225570"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.459145 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "cc385f93-2111-42cf-93be-d4cc82225570" (UID: "cc385f93-2111-42cf-93be-d4cc82225570"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.460823 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cc385f93-2111-42cf-93be-d4cc82225570" (UID: "cc385f93-2111-42cf-93be-d4cc82225570"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.465905 4946 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.465940 4946 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.465954 4946 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.465967 4946 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.465978 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.465989 4946 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cc385f93-2111-42cf-93be-d4cc82225570-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.466003 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sls7d\" (UniqueName: \"kubernetes.io/projected/cc385f93-2111-42cf-93be-d4cc82225570-kube-api-access-sls7d\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.758015 4946 generic.go:334] "Generic (PLEG): container finished" podID="cc385f93-2111-42cf-93be-d4cc82225570" containerID="a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be" exitCode=0 Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.758082 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86855dfc4c-jdcjc" event={"ID":"cc385f93-2111-42cf-93be-d4cc82225570","Type":"ContainerDied","Data":"a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be"} Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.758114 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86855dfc4c-jdcjc" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.758148 4946 scope.go:117] "RemoveContainer" containerID="01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.758129 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86855dfc4c-jdcjc" event={"ID":"cc385f93-2111-42cf-93be-d4cc82225570","Type":"ContainerDied","Data":"dc75f61ddd5966e8f97d174ed529f5c861cb0f2d892d432e30b91814e9ca814e"} Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.802993 4946 scope.go:117] "RemoveContainer" containerID="a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.806208 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86855dfc4c-jdcjc"] Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.818810 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-86855dfc4c-jdcjc"] Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.847452 4946 scope.go:117] "RemoveContainer" containerID="01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1" Dec 03 07:16:27 crc kubenswrapper[4946]: E1203 07:16:27.848094 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1\": container with ID starting with 01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1 not found: ID does not exist" containerID="01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.848143 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1"} err="failed to get container status \"01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1\": rpc error: code = NotFound desc = could not find container \"01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1\": container with ID starting with 01f1bd3078ac4b058a732e3e2bd019e9866a7c34748b7f63c646e100b80101f1 not found: ID does not exist" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.848177 4946 scope.go:117] "RemoveContainer" containerID="a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be" Dec 03 07:16:27 crc kubenswrapper[4946]: E1203 07:16:27.848967 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be\": container with ID starting with a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be not found: ID does not exist" containerID="a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be" Dec 03 07:16:27 crc kubenswrapper[4946]: I1203 07:16:27.849034 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be"} err="failed to get container status \"a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be\": rpc error: code = NotFound desc = could not find container \"a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be\": container with ID starting with a6d66ab256124a6141f035e21399696933099e235b1cc37dc0202614e43b05be not found: ID does not exist" Dec 03 07:16:29 crc kubenswrapper[4946]: I1203 07:16:29.609021 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc385f93-2111-42cf-93be-d4cc82225570" path="/var/lib/kubelet/pods/cc385f93-2111-42cf-93be-d4cc82225570/volumes" Dec 03 07:16:30 crc kubenswrapper[4946]: E1203 07:16:30.185664 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:30 crc kubenswrapper[4946]: E1203 07:16:30.186634 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:30 crc kubenswrapper[4946]: E1203 07:16:30.187195 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:30 crc kubenswrapper[4946]: E1203 07:16:30.187277 4946 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:30 crc kubenswrapper[4946]: E1203 07:16:30.187720 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:30 crc kubenswrapper[4946]: E1203 07:16:30.189891 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:30 crc kubenswrapper[4946]: E1203 07:16:30.193277 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:30 crc kubenswrapper[4946]: E1203 07:16:30.193329 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.600830 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.698214 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-combined-ca-bundle\") pod \"60cdd367-2981-42ad-a896-23ceee7ac34e\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.698301 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw8r7\" (UniqueName: \"kubernetes.io/projected/60cdd367-2981-42ad-a896-23ceee7ac34e-kube-api-access-bw8r7\") pod \"60cdd367-2981-42ad-a896-23ceee7ac34e\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.698380 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60cdd367-2981-42ad-a896-23ceee7ac34e-etc-machine-id\") pod \"60cdd367-2981-42ad-a896-23ceee7ac34e\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.698417 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data-custom\") pod \"60cdd367-2981-42ad-a896-23ceee7ac34e\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.698453 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-scripts\") pod \"60cdd367-2981-42ad-a896-23ceee7ac34e\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.698499 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data\") pod \"60cdd367-2981-42ad-a896-23ceee7ac34e\" (UID: \"60cdd367-2981-42ad-a896-23ceee7ac34e\") " Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.699121 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60cdd367-2981-42ad-a896-23ceee7ac34e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "60cdd367-2981-42ad-a896-23ceee7ac34e" (UID: "60cdd367-2981-42ad-a896-23ceee7ac34e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.706004 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "60cdd367-2981-42ad-a896-23ceee7ac34e" (UID: "60cdd367-2981-42ad-a896-23ceee7ac34e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.706022 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60cdd367-2981-42ad-a896-23ceee7ac34e-kube-api-access-bw8r7" (OuterVolumeSpecName: "kube-api-access-bw8r7") pod "60cdd367-2981-42ad-a896-23ceee7ac34e" (UID: "60cdd367-2981-42ad-a896-23ceee7ac34e"). InnerVolumeSpecName "kube-api-access-bw8r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.706071 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-scripts" (OuterVolumeSpecName: "scripts") pod "60cdd367-2981-42ad-a896-23ceee7ac34e" (UID: "60cdd367-2981-42ad-a896-23ceee7ac34e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.773288 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60cdd367-2981-42ad-a896-23ceee7ac34e" (UID: "60cdd367-2981-42ad-a896-23ceee7ac34e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.795775 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data" (OuterVolumeSpecName: "config-data") pod "60cdd367-2981-42ad-a896-23ceee7ac34e" (UID: "60cdd367-2981-42ad-a896-23ceee7ac34e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.801123 4946 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.801156 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw8r7\" (UniqueName: \"kubernetes.io/projected/60cdd367-2981-42ad-a896-23ceee7ac34e-kube-api-access-bw8r7\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.801169 4946 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60cdd367-2981-42ad-a896-23ceee7ac34e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.801178 4946 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.801188 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.801198 4946 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60cdd367-2981-42ad-a896-23ceee7ac34e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.839819 4946 generic.go:334] "Generic (PLEG): container finished" podID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerID="fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219" exitCode=137 Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.839863 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60cdd367-2981-42ad-a896-23ceee7ac34e","Type":"ContainerDied","Data":"fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219"} Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.839908 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"60cdd367-2981-42ad-a896-23ceee7ac34e","Type":"ContainerDied","Data":"42add7f84388be937cd02dd8af254dda6e580ef044a1f522802cffb491afe226"} Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.839890 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.839929 4946 scope.go:117] "RemoveContainer" containerID="ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.887766 4946 scope.go:117] "RemoveContainer" containerID="fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.913376 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.921174 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.922814 4946 scope.go:117] "RemoveContainer" containerID="ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374" Dec 03 07:16:33 crc kubenswrapper[4946]: E1203 07:16:33.923415 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374\": container with ID starting with ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374 not found: ID does not exist" containerID="ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.923445 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374"} err="failed to get container status \"ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374\": rpc error: code = NotFound desc = could not find container \"ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374\": container with ID starting with ea6ae1aec94f70bd98cc58b378ec076f61b83494c22e2fecc28e97a8c6ec5374 not found: ID does not exist" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.923466 4946 scope.go:117] "RemoveContainer" containerID="fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219" Dec 03 07:16:33 crc kubenswrapper[4946]: E1203 07:16:33.923840 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219\": container with ID starting with fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219 not found: ID does not exist" containerID="fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219" Dec 03 07:16:33 crc kubenswrapper[4946]: I1203 07:16:33.923869 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219"} err="failed to get container status \"fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219\": rpc error: code = NotFound desc = could not find container \"fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219\": container with ID starting with fccded26ebc187f5803f3b86ebc0c25ad1fa54dfc7e42eaddcdc4710dc764219 not found: ID does not exist" Dec 03 07:16:33 crc kubenswrapper[4946]: E1203 07:16:33.992118 4946 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60cdd367_2981_42ad_a896_23ceee7ac34e.slice\": RecentStats: unable to find data in memory cache]" Dec 03 07:16:34 crc kubenswrapper[4946]: I1203 07:16:34.593584 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:16:34 crc kubenswrapper[4946]: E1203 07:16:34.594127 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:16:35 crc kubenswrapper[4946]: E1203 07:16:35.189118 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:35 crc kubenswrapper[4946]: E1203 07:16:35.189320 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:35 crc kubenswrapper[4946]: E1203 07:16:35.190476 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:35 crc kubenswrapper[4946]: E1203 07:16:35.191032 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 03 07:16:35 crc kubenswrapper[4946]: E1203 07:16:35.191086 4946 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:35 crc kubenswrapper[4946]: E1203 07:16:35.192246 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:35 crc kubenswrapper[4946]: E1203 07:16:35.196052 4946 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 03 07:16:35 crc kubenswrapper[4946]: E1203 07:16:35.196162 4946 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sk56f" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.603471 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.611869 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" path="/var/lib/kubelet/pods/60cdd367-2981-42ad-a896-23ceee7ac34e/volumes" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.643644 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") pod \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.643772 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.643866 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-cache\") pod \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.643936 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-lock\") pod \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.644030 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l7qz\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-kube-api-access-5l7qz\") pod \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\" (UID: \"54cbc35f-afb5-4901-a64f-5e08fee6fd22\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.644445 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-cache" (OuterVolumeSpecName: "cache") pod "54cbc35f-afb5-4901-a64f-5e08fee6fd22" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.644734 4946 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-cache\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.645725 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-lock" (OuterVolumeSpecName: "lock") pod "54cbc35f-afb5-4901-a64f-5e08fee6fd22" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.656389 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "swift") pod "54cbc35f-afb5-4901-a64f-5e08fee6fd22" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.656617 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-kube-api-access-5l7qz" (OuterVolumeSpecName: "kube-api-access-5l7qz") pod "54cbc35f-afb5-4901-a64f-5e08fee6fd22" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22"). InnerVolumeSpecName "kube-api-access-5l7qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.667669 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "54cbc35f-afb5-4901-a64f-5e08fee6fd22" (UID: "54cbc35f-afb5-4901-a64f-5e08fee6fd22"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.748902 4946 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.748947 4946 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/54cbc35f-afb5-4901-a64f-5e08fee6fd22-lock\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.748965 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l7qz\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-kube-api-access-5l7qz\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.748989 4946 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/54cbc35f-afb5-4901-a64f-5e08fee6fd22-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.767585 4946 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.852694 4946 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.861227 4946 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod7b3d3641-40ca-4391-b7f2-811a85e9c99e"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod7b3d3641-40ca-4391-b7f2-811a85e9c99e] : Timed out while waiting for systemd to remove kubepods-besteffort-pod7b3d3641_40ca_4391_b7f2_811a85e9c99e.slice" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.871226 4946 generic.go:334] "Generic (PLEG): container finished" podID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerID="a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499" exitCode=137 Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.871319 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499"} Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.871358 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"54cbc35f-afb5-4901-a64f-5e08fee6fd22","Type":"ContainerDied","Data":"1793d3ed6844ec1e05f8d0259de7655d2a4b57adf23a1c3084e2c26c023d072d"} Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.871378 4946 scope.go:117] "RemoveContainer" containerID="a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.871665 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.873490 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sk56f_f8c1e670-a923-4bf5-a884-23644b2e52aa/ovs-vswitchd/0.log" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.875884 4946 generic.go:334] "Generic (PLEG): container finished" podID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" exitCode=137 Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.875914 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sk56f" event={"ID":"f8c1e670-a923-4bf5-a884-23644b2e52aa","Type":"ContainerDied","Data":"019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b"} Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.910502 4946 scope.go:117] "RemoveContainer" containerID="a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.910984 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.918003 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.923764 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sk56f_f8c1e670-a923-4bf5-a884-23644b2e52aa/ovs-vswitchd/0.log" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.924408 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.935614 4946 scope.go:117] "RemoveContainer" containerID="3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.955466 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-log\") pod \"f8c1e670-a923-4bf5-a884-23644b2e52aa\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.956018 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-etc-ovs\") pod \"f8c1e670-a923-4bf5-a884-23644b2e52aa\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.956646 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8c1e670-a923-4bf5-a884-23644b2e52aa-scripts\") pod \"f8c1e670-a923-4bf5-a884-23644b2e52aa\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.959092 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-run\") pod \"f8c1e670-a923-4bf5-a884-23644b2e52aa\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.959396 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6pnz\" (UniqueName: \"kubernetes.io/projected/f8c1e670-a923-4bf5-a884-23644b2e52aa-kube-api-access-t6pnz\") pod \"f8c1e670-a923-4bf5-a884-23644b2e52aa\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.955895 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-log" (OuterVolumeSpecName: "var-log") pod "f8c1e670-a923-4bf5-a884-23644b2e52aa" (UID: "f8c1e670-a923-4bf5-a884-23644b2e52aa"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.956413 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "f8c1e670-a923-4bf5-a884-23644b2e52aa" (UID: "f8c1e670-a923-4bf5-a884-23644b2e52aa"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.958959 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8c1e670-a923-4bf5-a884-23644b2e52aa-scripts" (OuterVolumeSpecName: "scripts") pod "f8c1e670-a923-4bf5-a884-23644b2e52aa" (UID: "f8c1e670-a923-4bf5-a884-23644b2e52aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.959310 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-run" (OuterVolumeSpecName: "var-run") pod "f8c1e670-a923-4bf5-a884-23644b2e52aa" (UID: "f8c1e670-a923-4bf5-a884-23644b2e52aa"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.960341 4946 scope.go:117] "RemoveContainer" containerID="4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.960711 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-lib" (OuterVolumeSpecName: "var-lib") pod "f8c1e670-a923-4bf5-a884-23644b2e52aa" (UID: "f8c1e670-a923-4bf5-a884-23644b2e52aa"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.960439 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-lib\") pod \"f8c1e670-a923-4bf5-a884-23644b2e52aa\" (UID: \"f8c1e670-a923-4bf5-a884-23644b2e52aa\") " Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.962803 4946 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.962817 4946 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-lib\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.962826 4946 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.962848 4946 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f8c1e670-a923-4bf5-a884-23644b2e52aa-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.962857 4946 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f8c1e670-a923-4bf5-a884-23644b2e52aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.963268 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c1e670-a923-4bf5-a884-23644b2e52aa-kube-api-access-t6pnz" (OuterVolumeSpecName: "kube-api-access-t6pnz") pod "f8c1e670-a923-4bf5-a884-23644b2e52aa" (UID: "f8c1e670-a923-4bf5-a884-23644b2e52aa"). InnerVolumeSpecName "kube-api-access-t6pnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:35 crc kubenswrapper[4946]: I1203 07:16:35.981596 4946 scope.go:117] "RemoveContainer" containerID="19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.000100 4946 scope.go:117] "RemoveContainer" containerID="caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.031978 4946 scope.go:117] "RemoveContainer" containerID="7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.051529 4946 scope.go:117] "RemoveContainer" containerID="a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.064449 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6pnz\" (UniqueName: \"kubernetes.io/projected/f8c1e670-a923-4bf5-a884-23644b2e52aa-kube-api-access-t6pnz\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.073015 4946 scope.go:117] "RemoveContainer" containerID="cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.097323 4946 scope.go:117] "RemoveContainer" containerID="51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.121142 4946 scope.go:117] "RemoveContainer" containerID="941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.153323 4946 scope.go:117] "RemoveContainer" containerID="8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.185531 4946 scope.go:117] "RemoveContainer" containerID="4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.204951 4946 scope.go:117] "RemoveContainer" containerID="eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.230516 4946 scope.go:117] "RemoveContainer" containerID="2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.267504 4946 scope.go:117] "RemoveContainer" containerID="a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.267823 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499\": container with ID starting with a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499 not found: ID does not exist" containerID="a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.267864 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499"} err="failed to get container status \"a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499\": rpc error: code = NotFound desc = could not find container \"a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499\": container with ID starting with a9120af63ab40dd24d6fdfe85399b9b5cdab47ed5d4dbe87f308ac08b3ac0499 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.267889 4946 scope.go:117] "RemoveContainer" containerID="a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.268211 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a\": container with ID starting with a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a not found: ID does not exist" containerID="a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.268239 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a"} err="failed to get container status \"a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a\": rpc error: code = NotFound desc = could not find container \"a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a\": container with ID starting with a72e5a5c3c8d8bb6b625e973a772407b0ec6d22802eae56ca1562f793dc1384a not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.268255 4946 scope.go:117] "RemoveContainer" containerID="3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.268522 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d\": container with ID starting with 3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d not found: ID does not exist" containerID="3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.268552 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d"} err="failed to get container status \"3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d\": rpc error: code = NotFound desc = could not find container \"3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d\": container with ID starting with 3744c4a6918d925b1d97b402f980bd36a31951ac71881d3e9a889e92bb41ff5d not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.268568 4946 scope.go:117] "RemoveContainer" containerID="4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.268793 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19\": container with ID starting with 4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19 not found: ID does not exist" containerID="4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.268817 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19"} err="failed to get container status \"4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19\": rpc error: code = NotFound desc = could not find container \"4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19\": container with ID starting with 4f2175259b2a873499fcc2ff80ec88e6e09a0179bc2b929f8898cb9b2628df19 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.268846 4946 scope.go:117] "RemoveContainer" containerID="19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.269083 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973\": container with ID starting with 19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973 not found: ID does not exist" containerID="19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.269149 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973"} err="failed to get container status \"19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973\": rpc error: code = NotFound desc = could not find container \"19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973\": container with ID starting with 19216c5517c757c2502764d2f962757289e5ffc1244d55a9363318e08a9b5973 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.269165 4946 scope.go:117] "RemoveContainer" containerID="caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.269529 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59\": container with ID starting with caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59 not found: ID does not exist" containerID="caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.269577 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59"} err="failed to get container status \"caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59\": rpc error: code = NotFound desc = could not find container \"caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59\": container with ID starting with caca34e9147036bb1e84f0b5db7e14586241c86958abc5f7579f221e9e7f9b59 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.269615 4946 scope.go:117] "RemoveContainer" containerID="7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.270431 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a\": container with ID starting with 7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a not found: ID does not exist" containerID="7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.270464 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a"} err="failed to get container status \"7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a\": rpc error: code = NotFound desc = could not find container \"7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a\": container with ID starting with 7529ad97c0b9b9d7a2fd2ec12b0da199be63b3a5ce3e841317739b380178e83a not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.270482 4946 scope.go:117] "RemoveContainer" containerID="a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.270763 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e\": container with ID starting with a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e not found: ID does not exist" containerID="a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.270785 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e"} err="failed to get container status \"a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e\": rpc error: code = NotFound desc = could not find container \"a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e\": container with ID starting with a3b694ea5e98d77e1d4e54f86fb4541d104a62f0dbb112b9821e3565c2d0256e not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.270799 4946 scope.go:117] "RemoveContainer" containerID="cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.271043 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2\": container with ID starting with cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2 not found: ID does not exist" containerID="cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.271070 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2"} err="failed to get container status \"cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2\": rpc error: code = NotFound desc = could not find container \"cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2\": container with ID starting with cf3a2f3f90ef51500ae25f1f7e2302db43fb58ee549327db8e5cb56a905345d2 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.271086 4946 scope.go:117] "RemoveContainer" containerID="51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.271429 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51\": container with ID starting with 51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51 not found: ID does not exist" containerID="51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.271451 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51"} err="failed to get container status \"51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51\": rpc error: code = NotFound desc = could not find container \"51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51\": container with ID starting with 51cc8a5f60470a99596490d060bbded767885a26c75188fa528f84947ef9ea51 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.271464 4946 scope.go:117] "RemoveContainer" containerID="941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.271781 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7\": container with ID starting with 941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7 not found: ID does not exist" containerID="941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.271807 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7"} err="failed to get container status \"941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7\": rpc error: code = NotFound desc = could not find container \"941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7\": container with ID starting with 941a1707e6ee3b2d85a2d2f95d73a243a4cc5e8df490cf461d548591e781fec7 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.271825 4946 scope.go:117] "RemoveContainer" containerID="8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.272206 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a\": container with ID starting with 8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a not found: ID does not exist" containerID="8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.272236 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a"} err="failed to get container status \"8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a\": rpc error: code = NotFound desc = could not find container \"8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a\": container with ID starting with 8ab4f627b61f3b27682b3b50c80db2f9634dee03ebe76b1cbbed0f2974bbef7a not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.272253 4946 scope.go:117] "RemoveContainer" containerID="4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.272569 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1\": container with ID starting with 4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1 not found: ID does not exist" containerID="4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.272620 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1"} err="failed to get container status \"4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1\": rpc error: code = NotFound desc = could not find container \"4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1\": container with ID starting with 4732830ace98de46786aa7e48103370e1cc129c59eb51f009793400b375901c1 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.272651 4946 scope.go:117] "RemoveContainer" containerID="eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.273134 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c\": container with ID starting with eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c not found: ID does not exist" containerID="eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.273166 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c"} err="failed to get container status \"eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c\": rpc error: code = NotFound desc = could not find container \"eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c\": container with ID starting with eef25065a31afb26e4828c480ea03765628a7107c5f28757d0666f0079bfa83c not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.273185 4946 scope.go:117] "RemoveContainer" containerID="2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826" Dec 03 07:16:36 crc kubenswrapper[4946]: E1203 07:16:36.273595 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826\": container with ID starting with 2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826 not found: ID does not exist" containerID="2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.273623 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826"} err="failed to get container status \"2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826\": rpc error: code = NotFound desc = could not find container \"2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826\": container with ID starting with 2cd358b49b10381505161380dac2f5e6272ea99c82c4fcfad3360db878520826 not found: ID does not exist" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.892476 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sk56f_f8c1e670-a923-4bf5-a884-23644b2e52aa/ovs-vswitchd/0.log" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.893662 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sk56f" event={"ID":"f8c1e670-a923-4bf5-a884-23644b2e52aa","Type":"ContainerDied","Data":"8191b4220a9e3c19d14db76a25889dc5cc3fdcb424688a3510ab0ba368e40f2c"} Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.893724 4946 scope.go:117] "RemoveContainer" containerID="019eebccb783af7718da03348c8def3f3d692a00e11c3eb5df5893f6a5f8101b" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.893862 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sk56f" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.924008 4946 scope.go:117] "RemoveContainer" containerID="2e9ab59f04b94e041a2b07f00c2b76480bfb2c697dc00955b425d0523765a732" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.954516 4946 scope.go:117] "RemoveContainer" containerID="b26d19402ccd1d5d3bfc46b1bae7e60dec0a258f0bebe9c9185a23fec83336f4" Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.956541 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-sk56f"] Dec 03 07:16:36 crc kubenswrapper[4946]: I1203 07:16:36.966522 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-sk56f"] Dec 03 07:16:37 crc kubenswrapper[4946]: I1203 07:16:37.608644 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" path="/var/lib/kubelet/pods/54cbc35f-afb5-4901-a64f-5e08fee6fd22/volumes" Dec 03 07:16:37 crc kubenswrapper[4946]: I1203 07:16:37.612625 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" path="/var/lib/kubelet/pods/f8c1e670-a923-4bf5-a884-23644b2e52aa/volumes" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.700659 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.705262 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.710182 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.716954 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.755202 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts\") pod \"572a50eb-42e9-4897-9925-4073b5f6a35d\" (UID: \"572a50eb-42e9-4897-9925-4073b5f6a35d\") " Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.755268 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ttpl\" (UniqueName: \"kubernetes.io/projected/85a6dcc4-23cc-4f57-8714-06d0ef25c680-kube-api-access-2ttpl\") pod \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\" (UID: \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\") " Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.755296 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wjw8\" (UniqueName: \"kubernetes.io/projected/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-kube-api-access-7wjw8\") pod \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\" (UID: \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\") " Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.755370 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts\") pod \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\" (UID: \"85a6dcc4-23cc-4f57-8714-06d0ef25c680\") " Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.755403 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5gxj\" (UniqueName: \"kubernetes.io/projected/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-kube-api-access-c5gxj\") pod \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\" (UID: \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\") " Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.755435 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts\") pod \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\" (UID: \"2f4c6455-fec3-41e2-bf0a-3595e34d63aa\") " Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.755462 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8pc7\" (UniqueName: \"kubernetes.io/projected/572a50eb-42e9-4897-9925-4073b5f6a35d-kube-api-access-c8pc7\") pod \"572a50eb-42e9-4897-9925-4073b5f6a35d\" (UID: \"572a50eb-42e9-4897-9925-4073b5f6a35d\") " Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.755487 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts\") pod \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\" (UID: \"b6e24ccc-b386-4d6e-9c67-cc95e47cafab\") " Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.756221 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85a6dcc4-23cc-4f57-8714-06d0ef25c680" (UID: "85a6dcc4-23cc-4f57-8714-06d0ef25c680"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.756254 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "572a50eb-42e9-4897-9925-4073b5f6a35d" (UID: "572a50eb-42e9-4897-9925-4073b5f6a35d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.756282 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6e24ccc-b386-4d6e-9c67-cc95e47cafab" (UID: "b6e24ccc-b386-4d6e-9c67-cc95e47cafab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.756817 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2f4c6455-fec3-41e2-bf0a-3595e34d63aa" (UID: "2f4c6455-fec3-41e2-bf0a-3595e34d63aa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.761149 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85a6dcc4-23cc-4f57-8714-06d0ef25c680-kube-api-access-2ttpl" (OuterVolumeSpecName: "kube-api-access-2ttpl") pod "85a6dcc4-23cc-4f57-8714-06d0ef25c680" (UID: "85a6dcc4-23cc-4f57-8714-06d0ef25c680"). InnerVolumeSpecName "kube-api-access-2ttpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.761218 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-kube-api-access-c5gxj" (OuterVolumeSpecName: "kube-api-access-c5gxj") pod "2f4c6455-fec3-41e2-bf0a-3595e34d63aa" (UID: "2f4c6455-fec3-41e2-bf0a-3595e34d63aa"). InnerVolumeSpecName "kube-api-access-c5gxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.761264 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572a50eb-42e9-4897-9925-4073b5f6a35d-kube-api-access-c8pc7" (OuterVolumeSpecName: "kube-api-access-c8pc7") pod "572a50eb-42e9-4897-9925-4073b5f6a35d" (UID: "572a50eb-42e9-4897-9925-4073b5f6a35d"). InnerVolumeSpecName "kube-api-access-c8pc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.761780 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-kube-api-access-7wjw8" (OuterVolumeSpecName: "kube-api-access-7wjw8") pod "b6e24ccc-b386-4d6e-9c67-cc95e47cafab" (UID: "b6e24ccc-b386-4d6e-9c67-cc95e47cafab"). InnerVolumeSpecName "kube-api-access-7wjw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.856983 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85a6dcc4-23cc-4f57-8714-06d0ef25c680-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.857008 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5gxj\" (UniqueName: \"kubernetes.io/projected/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-kube-api-access-c5gxj\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.857019 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4c6455-fec3-41e2-bf0a-3595e34d63aa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.857029 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8pc7\" (UniqueName: \"kubernetes.io/projected/572a50eb-42e9-4897-9925-4073b5f6a35d-kube-api-access-c8pc7\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.857038 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.857047 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/572a50eb-42e9-4897-9925-4073b5f6a35d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.857058 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ttpl\" (UniqueName: \"kubernetes.io/projected/85a6dcc4-23cc-4f57-8714-06d0ef25c680-kube-api-access-2ttpl\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.857067 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wjw8\" (UniqueName: \"kubernetes.io/projected/b6e24ccc-b386-4d6e-9c67-cc95e47cafab-kube-api-access-7wjw8\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.941213 4946 generic.go:334] "Generic (PLEG): container finished" podID="572a50eb-42e9-4897-9925-4073b5f6a35d" containerID="69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526" exitCode=137 Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.941285 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0960-account-delete-pc5x5" event={"ID":"572a50eb-42e9-4897-9925-4073b5f6a35d","Type":"ContainerDied","Data":"69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526"} Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.941290 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement0960-account-delete-pc5x5" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.941312 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0960-account-delete-pc5x5" event={"ID":"572a50eb-42e9-4897-9925-4073b5f6a35d","Type":"ContainerDied","Data":"1a891739c6bad23592f03f5fbaa0c50d11373da912e159446c140e5d10c4e675"} Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.941331 4946 scope.go:117] "RemoveContainer" containerID="69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.948756 4946 generic.go:334] "Generic (PLEG): container finished" podID="85a6dcc4-23cc-4f57-8714-06d0ef25c680" containerID="93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932" exitCode=137 Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.948810 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicaneae6-account-delete-l9467" event={"ID":"85a6dcc4-23cc-4f57-8714-06d0ef25c680","Type":"ContainerDied","Data":"93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932"} Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.948832 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicaneae6-account-delete-l9467" event={"ID":"85a6dcc4-23cc-4f57-8714-06d0ef25c680","Type":"ContainerDied","Data":"1e2834fdf3e1e0b071ee3d2989261ba331c5000d483d4e7b2af82af3bbc9b178"} Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.948886 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicaneae6-account-delete-l9467" Dec 03 07:16:40 crc kubenswrapper[4946]: E1203 07:16:40.959518 4946 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.959549 4946 generic.go:334] "Generic (PLEG): container finished" podID="0bc0555a-6c39-4e57-8c30-6a6fc422376a" containerID="0849c2b46e031c79a2d5d154130dc5d8f1d1b025f1f9369edf5dfdd4494048f0" exitCode=137 Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.959582 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0a5c7-account-delete-j2fhk" event={"ID":"0bc0555a-6c39-4e57-8c30-6a6fc422376a","Type":"ContainerDied","Data":"0849c2b46e031c79a2d5d154130dc5d8f1d1b025f1f9369edf5dfdd4494048f0"} Dec 03 07:16:40 crc kubenswrapper[4946]: E1203 07:16:40.959591 4946 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts podName:0bc0555a-6c39-4e57-8c30-6a6fc422376a nodeName:}" failed. No retries permitted until 2025-12-03 07:17:12.959573646 +0000 UTC m=+1625.756263765 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts") pod "novacell0a5c7-account-delete-j2fhk" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a") : configmap "openstack-scripts" not found Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.961197 4946 generic.go:334] "Generic (PLEG): container finished" podID="2f4c6455-fec3-41e2-bf0a-3595e34d63aa" containerID="11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d" exitCode=137 Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.961240 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi2e82-account-delete-77lgb" event={"ID":"2f4c6455-fec3-41e2-bf0a-3595e34d63aa","Type":"ContainerDied","Data":"11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d"} Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.961255 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi2e82-account-delete-77lgb" event={"ID":"2f4c6455-fec3-41e2-bf0a-3595e34d63aa","Type":"ContainerDied","Data":"c5dd963b92e27b3d88ed547b1a50b5d06468216e8e20c7b71c78967617c31a10"} Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.961288 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi2e82-account-delete-77lgb" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.966412 4946 generic.go:334] "Generic (PLEG): container finished" podID="b6e24ccc-b386-4d6e-9c67-cc95e47cafab" containerID="bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b" exitCode=137 Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.966463 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron6e16-account-delete-snvqr" event={"ID":"b6e24ccc-b386-4d6e-9c67-cc95e47cafab","Type":"ContainerDied","Data":"bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b"} Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.966496 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron6e16-account-delete-snvqr" event={"ID":"b6e24ccc-b386-4d6e-9c67-cc95e47cafab","Type":"ContainerDied","Data":"d6ca54dc63480a9b69f63634dde246e53accc0de903e525d1b3d24fce74ad4dd"} Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.966577 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron6e16-account-delete-snvqr" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.968796 4946 scope.go:117] "RemoveContainer" containerID="69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526" Dec 03 07:16:40 crc kubenswrapper[4946]: E1203 07:16:40.969194 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526\": container with ID starting with 69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526 not found: ID does not exist" containerID="69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.969241 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526"} err="failed to get container status \"69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526\": rpc error: code = NotFound desc = could not find container \"69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526\": container with ID starting with 69f191ec756079d2185780657ee9c0d9c05ee4019dd80f26db91833d5a3f5526 not found: ID does not exist" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.969284 4946 scope.go:117] "RemoveContainer" containerID="93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932" Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.988074 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicaneae6-account-delete-l9467"] Dec 03 07:16:40 crc kubenswrapper[4946]: I1203 07:16:40.995867 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicaneae6-account-delete-l9467"] Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.013533 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement0960-account-delete-pc5x5"] Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.014096 4946 scope.go:117] "RemoveContainer" containerID="93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932" Dec 03 07:16:41 crc kubenswrapper[4946]: E1203 07:16:41.016972 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932\": container with ID starting with 93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932 not found: ID does not exist" containerID="93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.017018 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932"} err="failed to get container status \"93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932\": rpc error: code = NotFound desc = could not find container \"93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932\": container with ID starting with 93748733c82b1a7d68c7bedf938f790b485ede86706160fa7dc30e25b2407932 not found: ID does not exist" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.017046 4946 scope.go:117] "RemoveContainer" containerID="11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.030686 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement0960-account-delete-pc5x5"] Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.034434 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.038399 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron6e16-account-delete-snvqr"] Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.049074 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron6e16-account-delete-snvqr"] Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.058331 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi2e82-account-delete-77lgb"] Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.059119 4946 scope.go:117] "RemoveContainer" containerID="11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d" Dec 03 07:16:41 crc kubenswrapper[4946]: E1203 07:16:41.059498 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d\": container with ID starting with 11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d not found: ID does not exist" containerID="11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.059547 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d"} err="failed to get container status \"11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d\": rpc error: code = NotFound desc = could not find container \"11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d\": container with ID starting with 11b29c8af1de02d3581d530fc6dc6bf3eed50f2e4f72f0449d9343424219a35d not found: ID does not exist" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.059598 4946 scope.go:117] "RemoveContainer" containerID="bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.060871 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi2e82-account-delete-77lgb"] Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.091414 4946 scope.go:117] "RemoveContainer" containerID="bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b" Dec 03 07:16:41 crc kubenswrapper[4946]: E1203 07:16:41.091727 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b\": container with ID starting with bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b not found: ID does not exist" containerID="bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.091786 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b"} err="failed to get container status \"bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b\": rpc error: code = NotFound desc = could not find container \"bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b\": container with ID starting with bfdde09f33cbefc801af3da88b025ddf16b8653ccde5e98bb377804dfd964f2b not found: ID does not exist" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.161475 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbpfg\" (UniqueName: \"kubernetes.io/projected/0bc0555a-6c39-4e57-8c30-6a6fc422376a-kube-api-access-nbpfg\") pod \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\" (UID: \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\") " Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.161616 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts\") pod \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\" (UID: \"0bc0555a-6c39-4e57-8c30-6a6fc422376a\") " Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.162294 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0bc0555a-6c39-4e57-8c30-6a6fc422376a" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.164197 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc0555a-6c39-4e57-8c30-6a6fc422376a-kube-api-access-nbpfg" (OuterVolumeSpecName: "kube-api-access-nbpfg") pod "0bc0555a-6c39-4e57-8c30-6a6fc422376a" (UID: "0bc0555a-6c39-4e57-8c30-6a6fc422376a"). InnerVolumeSpecName "kube-api-access-nbpfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.263581 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbpfg\" (UniqueName: \"kubernetes.io/projected/0bc0555a-6c39-4e57-8c30-6a6fc422376a-kube-api-access-nbpfg\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.263632 4946 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc0555a-6c39-4e57-8c30-6a6fc422376a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.612627 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f4c6455-fec3-41e2-bf0a-3595e34d63aa" path="/var/lib/kubelet/pods/2f4c6455-fec3-41e2-bf0a-3595e34d63aa/volumes" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.613800 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572a50eb-42e9-4897-9925-4073b5f6a35d" path="/var/lib/kubelet/pods/572a50eb-42e9-4897-9925-4073b5f6a35d/volumes" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.615203 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85a6dcc4-23cc-4f57-8714-06d0ef25c680" path="/var/lib/kubelet/pods/85a6dcc4-23cc-4f57-8714-06d0ef25c680/volumes" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.616333 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e24ccc-b386-4d6e-9c67-cc95e47cafab" path="/var/lib/kubelet/pods/b6e24ccc-b386-4d6e-9c67-cc95e47cafab/volumes" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.982706 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0a5c7-account-delete-j2fhk" event={"ID":"0bc0555a-6c39-4e57-8c30-6a6fc422376a","Type":"ContainerDied","Data":"32802bcbc5b55453efc9309d02e07fb693777d8deb061e6febf04a031e582b91"} Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.982802 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0a5c7-account-delete-j2fhk" Dec 03 07:16:41 crc kubenswrapper[4946]: I1203 07:16:41.983946 4946 scope.go:117] "RemoveContainer" containerID="0849c2b46e031c79a2d5d154130dc5d8f1d1b025f1f9369edf5dfdd4494048f0" Dec 03 07:16:42 crc kubenswrapper[4946]: I1203 07:16:42.019730 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0a5c7-account-delete-j2fhk"] Dec 03 07:16:42 crc kubenswrapper[4946]: I1203 07:16:42.030534 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0a5c7-account-delete-j2fhk"] Dec 03 07:16:43 crc kubenswrapper[4946]: I1203 07:16:43.609833 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bc0555a-6c39-4e57-8c30-6a6fc422376a" path="/var/lib/kubelet/pods/0bc0555a-6c39-4e57-8c30-6a6fc422376a/volumes" Dec 03 07:16:47 crc kubenswrapper[4946]: I1203 07:16:47.600447 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:16:47 crc kubenswrapper[4946]: E1203 07:16:47.601812 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.419900 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-flr4x"] Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421169 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerName="barbican-keystone-listener" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421193 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerName="barbican-keystone-listener" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421214 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" containerName="nova-cell1-conductor-conductor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421227 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" containerName="nova-cell1-conductor-conductor" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421246 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421258 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421284 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerName="cinder-scheduler" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421296 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerName="cinder-scheduler" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421315 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85593cc4-b6bd-44e9-a908-58f99f0c8756" containerName="galera" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421327 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="85593cc4-b6bd-44e9-a908-58f99f0c8756" containerName="galera" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421345 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421359 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421382 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421393 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-api" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421412 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-server" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421424 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-server" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421440 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.421451 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.421470 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85593cc4-b6bd-44e9-a908-58f99f0c8756" containerName="mysql-bootstrap" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424447 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="85593cc4-b6bd-44e9-a908-58f99f0c8756" containerName="mysql-bootstrap" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424507 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-updater" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424531 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-updater" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424561 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerName="barbican-worker" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424578 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerName="barbican-worker" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424610 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerName="barbican-keystone-listener-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424628 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerName="barbican-keystone-listener-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424653 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424669 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424704 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-reaper" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424726 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-reaper" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424795 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e57ed09-f7ce-403e-9cb5-676ae1c64117" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424816 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e57ed09-f7ce-403e-9cb5-676ae1c64117" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424846 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424863 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424897 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc0555a-6c39-4e57-8c30-6a6fc422376a" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424916 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc0555a-6c39-4e57-8c30-6a6fc422376a" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424948 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85a6dcc4-23cc-4f57-8714-06d0ef25c680" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.424968 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="85a6dcc4-23cc-4f57-8714-06d0ef25c680" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.424994 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f821218-d2a4-4f97-a924-199c056c7c55" containerName="rabbitmq" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425011 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f821218-d2a4-4f97-a924-199c056c7c55" containerName="rabbitmq" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425034 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425053 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425069 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425085 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425103 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="swift-recon-cron" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425119 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="swift-recon-cron" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425142 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerName="barbican-worker-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425158 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerName="barbican-worker-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425184 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="ceilometer-central-agent" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425201 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="ceilometer-central-agent" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425220 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425236 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425265 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="ovn-northd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425282 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="ovn-northd" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425300 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc385f93-2111-42cf-93be-d4cc82225570" containerName="neutron-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425316 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc385f93-2111-42cf-93be-d4cc82225570" containerName="neutron-api" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425346 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e0b4114-0121-48cc-9439-bc5d350aeb44" containerName="nova-cell0-conductor-conductor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425363 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e0b4114-0121-48cc-9439-bc5d350aeb44" containerName="nova-cell0-conductor-conductor" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425394 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="854575dc-d77d-43fe-a7ff-495ef76ab456" containerName="rabbitmq" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425410 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="854575dc-d77d-43fe-a7ff-495ef76ab456" containerName="rabbitmq" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425426 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5415d2e-cf73-4dd8-900d-211b957e829f" containerName="nova-scheduler-scheduler" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425445 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5415d2e-cf73-4dd8-900d-211b957e829f" containerName="nova-scheduler-scheduler" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425469 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-server" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425485 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-server" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425512 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerName="probe" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425527 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerName="probe" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425546 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-metadata" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425563 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-metadata" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425588 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425605 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425630 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="854575dc-d77d-43fe-a7ff-495ef76ab456" containerName="setup-container" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425645 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="854575dc-d77d-43fe-a7ff-495ef76ab456" containerName="setup-container" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425675 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server-init" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425694 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server-init" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425714 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425730 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425801 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e30f3ad7-043b-4212-a496-656c3447b509" containerName="keystone-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425821 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e30f3ad7-043b-4212-a496-656c3447b509" containerName="keystone-api" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425853 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="openstack-network-exporter" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425870 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="openstack-network-exporter" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425895 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-expirer" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425912 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-expirer" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425932 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425949 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.425975 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4c6455-fec3-41e2-bf0a-3595e34d63aa" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.425992 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4c6455-fec3-41e2-bf0a-3595e34d63aa" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426026 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426043 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426063 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="sg-core" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426083 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="sg-core" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426103 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426121 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-api" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426143 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf286a1-0b76-4e66-817c-7d7ce84edaa2" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426165 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf286a1-0b76-4e66-817c-7d7ce84edaa2" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426197 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426214 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426244 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerName="glance-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426261 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerName="glance-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426288 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69ee496-d325-40e0-81ff-0fb3625a775a" containerName="kube-state-metrics" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426307 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69ee496-d325-40e0-81ff-0fb3625a775a" containerName="kube-state-metrics" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426328 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426344 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426369 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerName="glance-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426385 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerName="glance-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426414 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerName="glance-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426430 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerName="glance-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426456 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426471 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426501 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="ceilometer-notification-agent" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426517 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="ceilometer-notification-agent" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426545 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerName="glance-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426562 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerName="glance-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426586 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f821218-d2a4-4f97-a924-199c056c7c55" containerName="setup-container" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426602 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f821218-d2a4-4f97-a924-199c056c7c55" containerName="setup-container" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426622 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572a50eb-42e9-4897-9925-4073b5f6a35d" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426639 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="572a50eb-42e9-4897-9925-4073b5f6a35d" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426663 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="rsync" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426678 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="rsync" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426703 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426784 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426815 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="proxy-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426831 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="proxy-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426850 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e24ccc-b386-4d6e-9c67-cc95e47cafab" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426866 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e24ccc-b386-4d6e-9c67-cc95e47cafab" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426889 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-server" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426906 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-server" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426936 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc385f93-2111-42cf-93be-d4cc82225570" containerName="neutron-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426951 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc385f93-2111-42cf-93be-d4cc82225570" containerName="neutron-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.426978 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36490b8-d5db-49c3-97e4-b734feaabd00" containerName="memcached" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.426995 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36490b8-d5db-49c3-97e4-b734feaabd00" containerName="memcached" Dec 03 07:16:55 crc kubenswrapper[4946]: E1203 07:16:55.427021 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-updater" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427035 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-updater" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427442 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427481 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427503 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="e30f3ad7-043b-4212-a496-656c3447b509" containerName="keystone-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427521 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36490b8-d5db-49c3-97e4-b734feaabd00" containerName="memcached" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427548 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerName="glance-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427571 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="ceilometer-notification-agent" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427596 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="ceilometer-central-agent" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427612 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5415d2e-cf73-4dd8-900d-211b957e829f" containerName="nova-scheduler-scheduler" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427636 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="85593cc4-b6bd-44e9-a908-58f99f0c8756" containerName="galera" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427661 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="rsync" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427691 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerName="probe" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427718 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="572a50eb-42e9-4897-9925-4073b5f6a35d" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427783 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cdd367-2981-42ad-a896-23ceee7ac34e" containerName="cinder-scheduler" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427801 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e24ccc-b386-4d6e-9c67-cc95e47cafab" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427825 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427841 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc0555a-6c39-4e57-8c30-6a6fc422376a" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427856 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f4c6455-fec3-41e2-bf0a-3595e34d63aa" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427878 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-expirer" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427895 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427915 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovsdb-server" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427932 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="761f2199-bf20-487a-a18d-172c65a4b515" containerName="barbican-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427952 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerName="glance-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.427980 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="openstack-network-exporter" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428005 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-server" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428027 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428045 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-updater" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428073 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="854575dc-d77d-43fe-a7ff-495ef76ab456" containerName="rabbitmq" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428097 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="proxy-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428118 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69ee496-d325-40e0-81ff-0fb3625a775a" containerName="kube-state-metrics" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428137 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-updater" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428156 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="85a6dcc4-23cc-4f57-8714-06d0ef25c680" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428177 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="be4f07e5-c606-4eb4-95cc-5e5c7dcb195d" containerName="ovn-controller" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428203 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="swift-recon-cron" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428219 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="10d9fdfc-0d03-4ca5-ba0f-2f8c408b55de" containerName="nova-cell1-conductor-conductor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428236 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerName="barbican-worker-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428254 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e57ed09-f7ce-403e-9cb5-676ae1c64117" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428274 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-server" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428290 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428318 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="80557a83-debb-4e3a-a2d0-bb5ac72d824c" containerName="ovn-northd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428335 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-auditor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428356 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc385f93-2111-42cf-93be-d4cc82225570" containerName="neutron-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428377 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428397 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerName="barbican-keystone-listener-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428415 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c1e670-a923-4bf5-a884-23644b2e52aa" containerName="ovs-vswitchd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428435 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="26ac177b-b889-43e7-888e-8ed02fb3ac72" containerName="barbican-worker" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428459 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428477 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd1d9c96-a0f4-43db-a686-2b320c1948ae" containerName="glance-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428494 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f821218-d2a4-4f97-a924-199c056c7c55" containerName="rabbitmq" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428516 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428535 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a3521a3-0f30-4afe-b5f8-efeb6f838e72" containerName="cinder-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428562 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf286a1-0b76-4e66-817c-7d7ce84edaa2" containerName="mariadb-account-delete" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428592 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ce5650-798f-4d12-bbd2-4ad5f2c47c5e" containerName="barbican-keystone-listener" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428656 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dc4eff7-f011-4265-b9dd-4e6bcdcb08df" containerName="nova-api-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428676 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="account-reaper" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428696 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d0798c-6a8a-4df3-9270-c75d49a2379c" containerName="sg-core" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428714 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="object-replicator" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428777 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbc35f-afb5-4901-a64f-5e08fee6fd22" containerName="container-server" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428797 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3ead5e-98d3-4719-b1d2-cda4d7b7b541" containerName="placement-log" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428824 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="192c2964-d6c9-404e-8cca-cac01c750c5b" containerName="nova-metadata-metadata" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428843 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc385f93-2111-42cf-93be-d4cc82225570" containerName="neutron-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428867 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7758850c-e98d-4bb6-bd84-31584a9bcdb6" containerName="glance-httpd" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.428893 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e0b4114-0121-48cc-9439-bc5d350aeb44" containerName="nova-cell0-conductor-conductor" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.431099 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.434002 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-flr4x"] Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.508169 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-utilities\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.508257 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmk7n\" (UniqueName: \"kubernetes.io/projected/0c3dddac-888c-49a2-b664-c9633b8365d9-kube-api-access-wmk7n\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.508373 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-catalog-content\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.609326 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-catalog-content\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.610051 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-catalog-content\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.611214 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-utilities\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.611332 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmk7n\" (UniqueName: \"kubernetes.io/projected/0c3dddac-888c-49a2-b664-c9633b8365d9-kube-api-access-wmk7n\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.612481 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-utilities\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.639486 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmk7n\" (UniqueName: \"kubernetes.io/projected/0c3dddac-888c-49a2-b664-c9633b8365d9-kube-api-access-wmk7n\") pod \"community-operators-flr4x\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:55 crc kubenswrapper[4946]: I1203 07:16:55.784841 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:16:56 crc kubenswrapper[4946]: I1203 07:16:56.339533 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-flr4x"] Dec 03 07:16:57 crc kubenswrapper[4946]: I1203 07:16:57.203061 4946 generic.go:334] "Generic (PLEG): container finished" podID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerID="9d4a4851bf3e0e74f9aa339ba98b2cba60131a95a5acd6eb3df6241cc29dd57b" exitCode=0 Dec 03 07:16:57 crc kubenswrapper[4946]: I1203 07:16:57.203138 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flr4x" event={"ID":"0c3dddac-888c-49a2-b664-c9633b8365d9","Type":"ContainerDied","Data":"9d4a4851bf3e0e74f9aa339ba98b2cba60131a95a5acd6eb3df6241cc29dd57b"} Dec 03 07:16:57 crc kubenswrapper[4946]: I1203 07:16:57.205059 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flr4x" event={"ID":"0c3dddac-888c-49a2-b664-c9633b8365d9","Type":"ContainerStarted","Data":"4cecc4223331a3f3c0dace82d53e12bf809dddd74120cfcd5a63759a573c9ad8"} Dec 03 07:16:58 crc kubenswrapper[4946]: I1203 07:16:58.217248 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flr4x" event={"ID":"0c3dddac-888c-49a2-b664-c9633b8365d9","Type":"ContainerStarted","Data":"fbe89bda3aca56604148534b93293581c804ded1704dcada07d727b576f65a94"} Dec 03 07:16:58 crc kubenswrapper[4946]: I1203 07:16:58.592961 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:16:58 crc kubenswrapper[4946]: E1203 07:16:58.593446 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:16:59 crc kubenswrapper[4946]: I1203 07:16:59.227678 4946 generic.go:334] "Generic (PLEG): container finished" podID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerID="fbe89bda3aca56604148534b93293581c804ded1704dcada07d727b576f65a94" exitCode=0 Dec 03 07:16:59 crc kubenswrapper[4946]: I1203 07:16:59.227722 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flr4x" event={"ID":"0c3dddac-888c-49a2-b664-c9633b8365d9","Type":"ContainerDied","Data":"fbe89bda3aca56604148534b93293581c804ded1704dcada07d727b576f65a94"} Dec 03 07:17:00 crc kubenswrapper[4946]: I1203 07:17:00.250455 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flr4x" event={"ID":"0c3dddac-888c-49a2-b664-c9633b8365d9","Type":"ContainerStarted","Data":"002b6804998f7178b62136dc035c899098680bd686d5cef439e205e9711b9112"} Dec 03 07:17:00 crc kubenswrapper[4946]: I1203 07:17:00.282342 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-flr4x" podStartSLOduration=2.47907222 podStartE2EDuration="5.282308294s" podCreationTimestamp="2025-12-03 07:16:55 +0000 UTC" firstStartedPulling="2025-12-03 07:16:57.205227741 +0000 UTC m=+1610.001917880" lastFinishedPulling="2025-12-03 07:17:00.008463805 +0000 UTC m=+1612.805153954" observedRunningTime="2025-12-03 07:17:00.27956032 +0000 UTC m=+1613.076250499" watchObservedRunningTime="2025-12-03 07:17:00.282308294 +0000 UTC m=+1613.078998473" Dec 03 07:17:05 crc kubenswrapper[4946]: I1203 07:17:05.785525 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:17:05 crc kubenswrapper[4946]: I1203 07:17:05.786380 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:17:05 crc kubenswrapper[4946]: I1203 07:17:05.866162 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:17:06 crc kubenswrapper[4946]: I1203 07:17:06.405442 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:17:06 crc kubenswrapper[4946]: I1203 07:17:06.480238 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-flr4x"] Dec 03 07:17:08 crc kubenswrapper[4946]: I1203 07:17:08.352702 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-flr4x" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerName="registry-server" containerID="cri-o://002b6804998f7178b62136dc035c899098680bd686d5cef439e205e9711b9112" gracePeriod=2 Dec 03 07:17:09 crc kubenswrapper[4946]: I1203 07:17:09.368288 4946 generic.go:334] "Generic (PLEG): container finished" podID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerID="002b6804998f7178b62136dc035c899098680bd686d5cef439e205e9711b9112" exitCode=0 Dec 03 07:17:09 crc kubenswrapper[4946]: I1203 07:17:09.368352 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flr4x" event={"ID":"0c3dddac-888c-49a2-b664-c9633b8365d9","Type":"ContainerDied","Data":"002b6804998f7178b62136dc035c899098680bd686d5cef439e205e9711b9112"} Dec 03 07:17:09 crc kubenswrapper[4946]: I1203 07:17:09.973423 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.150952 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-catalog-content\") pod \"0c3dddac-888c-49a2-b664-c9633b8365d9\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.151130 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmk7n\" (UniqueName: \"kubernetes.io/projected/0c3dddac-888c-49a2-b664-c9633b8365d9-kube-api-access-wmk7n\") pod \"0c3dddac-888c-49a2-b664-c9633b8365d9\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.151231 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-utilities\") pod \"0c3dddac-888c-49a2-b664-c9633b8365d9\" (UID: \"0c3dddac-888c-49a2-b664-c9633b8365d9\") " Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.152442 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-utilities" (OuterVolumeSpecName: "utilities") pod "0c3dddac-888c-49a2-b664-c9633b8365d9" (UID: "0c3dddac-888c-49a2-b664-c9633b8365d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.159279 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c3dddac-888c-49a2-b664-c9633b8365d9-kube-api-access-wmk7n" (OuterVolumeSpecName: "kube-api-access-wmk7n") pod "0c3dddac-888c-49a2-b664-c9633b8365d9" (UID: "0c3dddac-888c-49a2-b664-c9633b8365d9"). InnerVolumeSpecName "kube-api-access-wmk7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.205278 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c3dddac-888c-49a2-b664-c9633b8365d9" (UID: "0c3dddac-888c-49a2-b664-c9633b8365d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.253094 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.253159 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c3dddac-888c-49a2-b664-c9633b8365d9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.253174 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmk7n\" (UniqueName: \"kubernetes.io/projected/0c3dddac-888c-49a2-b664-c9633b8365d9-kube-api-access-wmk7n\") on node \"crc\" DevicePath \"\"" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.407403 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flr4x" event={"ID":"0c3dddac-888c-49a2-b664-c9633b8365d9","Type":"ContainerDied","Data":"4cecc4223331a3f3c0dace82d53e12bf809dddd74120cfcd5a63759a573c9ad8"} Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.407490 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flr4x" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.407545 4946 scope.go:117] "RemoveContainer" containerID="002b6804998f7178b62136dc035c899098680bd686d5cef439e205e9711b9112" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.437637 4946 scope.go:117] "RemoveContainer" containerID="fbe89bda3aca56604148534b93293581c804ded1704dcada07d727b576f65a94" Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.469213 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-flr4x"] Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.481937 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-flr4x"] Dec 03 07:17:10 crc kubenswrapper[4946]: I1203 07:17:10.499324 4946 scope.go:117] "RemoveContainer" containerID="9d4a4851bf3e0e74f9aa339ba98b2cba60131a95a5acd6eb3df6241cc29dd57b" Dec 03 07:17:11 crc kubenswrapper[4946]: I1203 07:17:11.609658 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" path="/var/lib/kubelet/pods/0c3dddac-888c-49a2-b664-c9633b8365d9/volumes" Dec 03 07:17:13 crc kubenswrapper[4946]: I1203 07:17:13.593265 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:17:13 crc kubenswrapper[4946]: E1203 07:17:13.593662 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:17:24 crc kubenswrapper[4946]: I1203 07:17:24.592938 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:17:24 crc kubenswrapper[4946]: E1203 07:17:24.594044 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:17:37 crc kubenswrapper[4946]: I1203 07:17:37.598214 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:17:37 crc kubenswrapper[4946]: E1203 07:17:37.601237 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:17:49 crc kubenswrapper[4946]: I1203 07:17:49.593118 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:17:49 crc kubenswrapper[4946]: E1203 07:17:49.593987 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:18:00 crc kubenswrapper[4946]: I1203 07:18:00.593089 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:18:00 crc kubenswrapper[4946]: E1203 07:18:00.596056 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.059606 4946 scope.go:117] "RemoveContainer" containerID="2d1d6e59deb3691d32b28325455c408f0366e63c827b8f379c3c4aae9e05ae72" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.106293 4946 scope.go:117] "RemoveContainer" containerID="6f7b1944bbb54c5a21449cb0e33457d34fcb8a0ab8e6a2a481e6fb4b73ed96cf" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.168057 4946 scope.go:117] "RemoveContainer" containerID="1108f67de44b6b63d6dab499b1df82ca4a27a61bd91177dc82437eaf17006c2a" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.200404 4946 scope.go:117] "RemoveContainer" containerID="a4ff77b363d0d7b530a95ecccec7f1113feeb3d0a4ee6fc234f8f36f96f15956" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.254766 4946 scope.go:117] "RemoveContainer" containerID="d697ca9b1e80be206a399092b24806b7f6fa0b97b98856c8b392319cdac0c1d4" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.294189 4946 scope.go:117] "RemoveContainer" containerID="d20e227fd20281f8afac7a5f2b06e1522069ec70539b7c99d3e1814e068f79e7" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.335700 4946 scope.go:117] "RemoveContainer" containerID="740fb16d62aafc273f14ed8ada7248138673e6e2b5d0bfcb331e6f2fed1a3dde" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.366184 4946 scope.go:117] "RemoveContainer" containerID="1c6942a4964ff011f829ade85e97de580a3421d534d44522ff6248ad37aa57cb" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.396461 4946 scope.go:117] "RemoveContainer" containerID="4ce9f26033ad21dcb676e42f79ece97ce0b09051bf144ebdb9aea2ae4003da81" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.419403 4946 scope.go:117] "RemoveContainer" containerID="6700c2fcb08611e2d7bdcdf960bb02fa23e8b4f20b45767e6216ee45698f3fc5" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.459412 4946 scope.go:117] "RemoveContainer" containerID="87aa8a68b0871fd821afa7b8d5609c2fd2858545673a2ba5c62eede8cb96b325" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.494405 4946 scope.go:117] "RemoveContainer" containerID="f3336a95769c42e445e00fea4ac613e0e2b8245b491ca7999ea149e6287a6f15" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.523345 4946 scope.go:117] "RemoveContainer" containerID="8d2fc8922be755e466d97af0830a141219adb403d1d1638734857140a464dd7a" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.543320 4946 scope.go:117] "RemoveContainer" containerID="d346ebb93b5866ea909e69f58ddb68b815c76d2379fcd0a1e7f521aa60a3c3fb" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.581499 4946 scope.go:117] "RemoveContainer" containerID="709c3aea255550cce7f8b4bca9af001a3c96f7d77f9a4aeadadf2f13c22bb5b5" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.644160 4946 scope.go:117] "RemoveContainer" containerID="a4c5ab9a4b7288722bd3503516f639271ce0628e104cf87e262381e3ac10cc21" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.675613 4946 scope.go:117] "RemoveContainer" containerID="8dcad6ca973112b601de433a0add9cb4252be0e542acfd5c99157c13bbe11c5a" Dec 03 07:18:10 crc kubenswrapper[4946]: I1203 07:18:10.744957 4946 scope.go:117] "RemoveContainer" containerID="243d3ab281717f39767cc58f3cd2061d2e057089fdd1678c26dadc9e51f4de2d" Dec 03 07:18:12 crc kubenswrapper[4946]: I1203 07:18:12.594445 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:18:12 crc kubenswrapper[4946]: E1203 07:18:12.595180 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:18:24 crc kubenswrapper[4946]: I1203 07:18:24.593143 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:18:24 crc kubenswrapper[4946]: E1203 07:18:24.594219 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:18:37 crc kubenswrapper[4946]: I1203 07:18:37.600901 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:18:37 crc kubenswrapper[4946]: E1203 07:18:37.602108 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:18:49 crc kubenswrapper[4946]: I1203 07:18:49.593643 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:18:49 crc kubenswrapper[4946]: E1203 07:18:49.594632 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:19:02 crc kubenswrapper[4946]: I1203 07:19:02.592861 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:19:02 crc kubenswrapper[4946]: E1203 07:19:02.593971 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:19:11 crc kubenswrapper[4946]: I1203 07:19:11.218893 4946 scope.go:117] "RemoveContainer" containerID="6d8145ae8f7303d9ebe9697e32c04cdfb3a46dd8ce1dc51d2f81b3c03b54e2c8" Dec 03 07:19:11 crc kubenswrapper[4946]: I1203 07:19:11.253641 4946 scope.go:117] "RemoveContainer" containerID="2c6c7a3b7db447b64ab670632d0dd83df47c93d671ed2f91a26e1146fa3b06ad" Dec 03 07:19:11 crc kubenswrapper[4946]: I1203 07:19:11.297603 4946 scope.go:117] "RemoveContainer" containerID="9cccbf44990a5cf20986b72a9645414dda943c7dde99da340999a5cf27c3cfe5" Dec 03 07:19:11 crc kubenswrapper[4946]: I1203 07:19:11.362271 4946 scope.go:117] "RemoveContainer" containerID="7d1e3a0daf343eed99376edb4c8bfb01cc40b826d5dc69f831d706bc8864bb79" Dec 03 07:19:11 crc kubenswrapper[4946]: I1203 07:19:11.435523 4946 scope.go:117] "RemoveContainer" containerID="0c1e1e5afa702c1d51ea8d3cbdd531ef58625215a8edb86478cd68e3ae7a611f" Dec 03 07:19:11 crc kubenswrapper[4946]: I1203 07:19:11.465775 4946 scope.go:117] "RemoveContainer" containerID="06396dccbda6437120d3bc95eb233be6e528a4975537c272e4321c4e7fb51398" Dec 03 07:19:15 crc kubenswrapper[4946]: I1203 07:19:15.592650 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:19:15 crc kubenswrapper[4946]: E1203 07:19:15.593476 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:19:27 crc kubenswrapper[4946]: I1203 07:19:27.600886 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:19:27 crc kubenswrapper[4946]: E1203 07:19:27.601959 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:19:38 crc kubenswrapper[4946]: I1203 07:19:38.592762 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:19:38 crc kubenswrapper[4946]: E1203 07:19:38.593718 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:19:53 crc kubenswrapper[4946]: I1203 07:19:53.593171 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:19:53 crc kubenswrapper[4946]: E1203 07:19:53.595839 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:20:04 crc kubenswrapper[4946]: I1203 07:20:04.592479 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:20:04 crc kubenswrapper[4946]: E1203 07:20:04.593654 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.628732 4946 scope.go:117] "RemoveContainer" containerID="a8fcde969d6587f94b75052c2082af4c747c7ae565921025cb63af54faad0ca2" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.661878 4946 scope.go:117] "RemoveContainer" containerID="8bbf7ed4bfd39c8452df4287d669a6ba62b2a15a4a2fbbe3cb29b95583fbc7ff" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.701914 4946 scope.go:117] "RemoveContainer" containerID="985f669d8f4553a11987acfa7117c970727e75f62281c757c43798f37b6ae8b7" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.767603 4946 scope.go:117] "RemoveContainer" containerID="02e3cd254b809591216a604bc32c72b24c11ab31b407688852a7678cf83aaefd" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.797047 4946 scope.go:117] "RemoveContainer" containerID="3094708632df5b159d2bf6133a4f3ee720f6efc93a22e2b6ef2a2b1ca81637fd" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.826502 4946 scope.go:117] "RemoveContainer" containerID="072f03744aa756782573999562fa06b8b69eba687793e806dfa9783e867eb97e" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.854374 4946 scope.go:117] "RemoveContainer" containerID="29cdfda04c231794a246ad8f35d301624ae9c4ccf115d7201fda3eefc9f74be9" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.888343 4946 scope.go:117] "RemoveContainer" containerID="bdd42e0495c25db0bfd7a6bee417a5e00de9e626919b6c07b7165d2e59deff1e" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.915809 4946 scope.go:117] "RemoveContainer" containerID="86f69893a1328049356c59c5662729d8162a84f1a36a98ea3bea5e219a4226d6" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.935331 4946 scope.go:117] "RemoveContainer" containerID="315daa8d7809bce784559041f6d32310303fe1c049ed5f9e01eed7bf24ba6320" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.966081 4946 scope.go:117] "RemoveContainer" containerID="16f4c8fc003db664d5e0f306fcdc880d7d0d495455f0d0af1082b3a8f6861d8c" Dec 03 07:20:11 crc kubenswrapper[4946]: I1203 07:20:11.992718 4946 scope.go:117] "RemoveContainer" containerID="0dacc1b75220a156eb9ecf0c38e75527ad903c7b827431881e01db50c9cbe8e1" Dec 03 07:20:12 crc kubenswrapper[4946]: I1203 07:20:12.022568 4946 scope.go:117] "RemoveContainer" containerID="fe14b127e3c75b219b08fa364540cfc92ff3fbff0bfffa175bcfa8f37f424f19" Dec 03 07:20:12 crc kubenswrapper[4946]: I1203 07:20:12.055352 4946 scope.go:117] "RemoveContainer" containerID="d8b6ba36f5630c14206454eeddd3f93c889a0334d0f55d1e24da48d1af5af11e" Dec 03 07:20:16 crc kubenswrapper[4946]: I1203 07:20:16.593872 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:20:16 crc kubenswrapper[4946]: E1203 07:20:16.595121 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:20:27 crc kubenswrapper[4946]: I1203 07:20:27.610627 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:20:27 crc kubenswrapper[4946]: E1203 07:20:27.611993 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:20:41 crc kubenswrapper[4946]: I1203 07:20:41.592867 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:20:41 crc kubenswrapper[4946]: E1203 07:20:41.594045 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:20:54 crc kubenswrapper[4946]: I1203 07:20:54.593184 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:20:54 crc kubenswrapper[4946]: E1203 07:20:54.594192 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:21:06 crc kubenswrapper[4946]: I1203 07:21:06.593787 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:21:06 crc kubenswrapper[4946]: E1203 07:21:06.594536 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:21:12 crc kubenswrapper[4946]: I1203 07:21:12.223076 4946 scope.go:117] "RemoveContainer" containerID="c38091b17a25c86f76292d47faafe82eaf48ed142e05937e9350d421568f4d36" Dec 03 07:21:12 crc kubenswrapper[4946]: I1203 07:21:12.275990 4946 scope.go:117] "RemoveContainer" containerID="b4d0c049e4ebf0ba2b784151e0a28518ad34dc7f7cabb79dbe4c2f1ad0f9a64a" Dec 03 07:21:12 crc kubenswrapper[4946]: I1203 07:21:12.322426 4946 scope.go:117] "RemoveContainer" containerID="2053513fb2cd98af806d2471e5ac121d3b4f53386b03f6ab961cfeb726f810cd" Dec 03 07:21:12 crc kubenswrapper[4946]: I1203 07:21:12.373131 4946 scope.go:117] "RemoveContainer" containerID="c6a3e5b8e2f5447abcb8ac293e3ff5788a44666315066833a7c030a378495ed9" Dec 03 07:21:12 crc kubenswrapper[4946]: I1203 07:21:12.404348 4946 scope.go:117] "RemoveContainer" containerID="d01f7981c5efe052b4cf6101ea504e6ce8ea707e982a9081e7754f677063658e" Dec 03 07:21:12 crc kubenswrapper[4946]: I1203 07:21:12.432361 4946 scope.go:117] "RemoveContainer" containerID="39788eaaa331a03bc7d5fe3fa41f1ea2143ebafa1975eab3dca5fb0ffc7bcf07" Dec 03 07:21:12 crc kubenswrapper[4946]: I1203 07:21:12.501026 4946 scope.go:117] "RemoveContainer" containerID="472b894497c195453df59fead8c87fdb10a93680b5df08ccc87ddf9ae8f84460" Dec 03 07:21:12 crc kubenswrapper[4946]: I1203 07:21:12.520384 4946 scope.go:117] "RemoveContainer" containerID="5b2360ca28716b38f9b4de81b0cdae9c770e347cdcfea1c775c2365830f4ebbf" Dec 03 07:21:18 crc kubenswrapper[4946]: I1203 07:21:18.593054 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:21:18 crc kubenswrapper[4946]: E1203 07:21:18.593978 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:21:29 crc kubenswrapper[4946]: I1203 07:21:29.594268 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:21:30 crc kubenswrapper[4946]: I1203 07:21:30.516666 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"f63f6469fc21ed7b131500e8679e2fed6eb0e75d27f7ea70e25254efa4de3f73"} Dec 03 07:22:12 crc kubenswrapper[4946]: I1203 07:22:12.653043 4946 scope.go:117] "RemoveContainer" containerID="cccc8d922dd34c8481eefd183a382905de547dc11a0e346c9e3f2854da1c3218" Dec 03 07:22:12 crc kubenswrapper[4946]: I1203 07:22:12.695350 4946 scope.go:117] "RemoveContainer" containerID="5f4981a54547ee9f0ce3adc448f87d4c7e86f80240ae2ecacd9efac2c2bb4710" Dec 03 07:23:53 crc kubenswrapper[4946]: I1203 07:23:53.039121 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:23:53 crc kubenswrapper[4946]: I1203 07:23:53.040025 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:24:23 crc kubenswrapper[4946]: I1203 07:24:23.040010 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:24:23 crc kubenswrapper[4946]: I1203 07:24:23.040711 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.714375 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nx9hz"] Dec 03 07:24:49 crc kubenswrapper[4946]: E1203 07:24:49.716692 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerName="extract-utilities" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.716708 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerName="extract-utilities" Dec 03 07:24:49 crc kubenswrapper[4946]: E1203 07:24:49.716768 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerName="registry-server" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.716776 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerName="registry-server" Dec 03 07:24:49 crc kubenswrapper[4946]: E1203 07:24:49.716787 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerName="extract-content" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.716794 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerName="extract-content" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.716921 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c3dddac-888c-49a2-b664-c9633b8365d9" containerName="registry-server" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.717937 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.725913 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nx9hz"] Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.806358 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-catalog-content\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.806441 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-945xc\" (UniqueName: \"kubernetes.io/projected/c10cd69d-808b-4196-b8c2-654f386beff6-kube-api-access-945xc\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.806518 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-utilities\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.907822 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-catalog-content\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.907906 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-945xc\" (UniqueName: \"kubernetes.io/projected/c10cd69d-808b-4196-b8c2-654f386beff6-kube-api-access-945xc\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.907977 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-utilities\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.908406 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-catalog-content\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.908443 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-utilities\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:49 crc kubenswrapper[4946]: I1203 07:24:49.926503 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-945xc\" (UniqueName: \"kubernetes.io/projected/c10cd69d-808b-4196-b8c2-654f386beff6-kube-api-access-945xc\") pod \"certified-operators-nx9hz\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:50 crc kubenswrapper[4946]: I1203 07:24:50.046301 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:24:50 crc kubenswrapper[4946]: I1203 07:24:50.489594 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nx9hz"] Dec 03 07:24:51 crc kubenswrapper[4946]: I1203 07:24:51.371780 4946 generic.go:334] "Generic (PLEG): container finished" podID="c10cd69d-808b-4196-b8c2-654f386beff6" containerID="157e3638c78e60c49818aa047d144e14f3c84423af0eab47b694a7e2c06e62ea" exitCode=0 Dec 03 07:24:51 crc kubenswrapper[4946]: I1203 07:24:51.371984 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx9hz" event={"ID":"c10cd69d-808b-4196-b8c2-654f386beff6","Type":"ContainerDied","Data":"157e3638c78e60c49818aa047d144e14f3c84423af0eab47b694a7e2c06e62ea"} Dec 03 07:24:51 crc kubenswrapper[4946]: I1203 07:24:51.372563 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx9hz" event={"ID":"c10cd69d-808b-4196-b8c2-654f386beff6","Type":"ContainerStarted","Data":"009315ec4e3f97bf91463a8d387aa739551abcca2908d0a1766fbfe3e79be59c"} Dec 03 07:24:51 crc kubenswrapper[4946]: I1203 07:24:51.376583 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:24:52 crc kubenswrapper[4946]: I1203 07:24:52.380460 4946 generic.go:334] "Generic (PLEG): container finished" podID="c10cd69d-808b-4196-b8c2-654f386beff6" containerID="3a0c462461d34e4d3b7768e77aa6fe2340dc1f628adee71ec15c82dd203fa18a" exitCode=0 Dec 03 07:24:52 crc kubenswrapper[4946]: I1203 07:24:52.380520 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx9hz" event={"ID":"c10cd69d-808b-4196-b8c2-654f386beff6","Type":"ContainerDied","Data":"3a0c462461d34e4d3b7768e77aa6fe2340dc1f628adee71ec15c82dd203fa18a"} Dec 03 07:24:53 crc kubenswrapper[4946]: I1203 07:24:53.039953 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:24:53 crc kubenswrapper[4946]: I1203 07:24:53.040011 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:24:53 crc kubenswrapper[4946]: I1203 07:24:53.040054 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:24:53 crc kubenswrapper[4946]: I1203 07:24:53.040642 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f63f6469fc21ed7b131500e8679e2fed6eb0e75d27f7ea70e25254efa4de3f73"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:24:53 crc kubenswrapper[4946]: I1203 07:24:53.040703 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://f63f6469fc21ed7b131500e8679e2fed6eb0e75d27f7ea70e25254efa4de3f73" gracePeriod=600 Dec 03 07:24:54 crc kubenswrapper[4946]: I1203 07:24:54.399700 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx9hz" event={"ID":"c10cd69d-808b-4196-b8c2-654f386beff6","Type":"ContainerStarted","Data":"654117bb1bbe9ea808dcb95f3fbf5f699dd2815df05e00fd44ab638e6da9a1e9"} Dec 03 07:24:54 crc kubenswrapper[4946]: I1203 07:24:54.403613 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="f63f6469fc21ed7b131500e8679e2fed6eb0e75d27f7ea70e25254efa4de3f73" exitCode=0 Dec 03 07:24:54 crc kubenswrapper[4946]: I1203 07:24:54.403662 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"f63f6469fc21ed7b131500e8679e2fed6eb0e75d27f7ea70e25254efa4de3f73"} Dec 03 07:24:54 crc kubenswrapper[4946]: I1203 07:24:54.403693 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad"} Dec 03 07:24:54 crc kubenswrapper[4946]: I1203 07:24:54.403718 4946 scope.go:117] "RemoveContainer" containerID="bd63a133f2ca5101c00daec98defe7545ba9349b12089d7bfd8db35e3ff6b113" Dec 03 07:24:54 crc kubenswrapper[4946]: I1203 07:24:54.431773 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nx9hz" podStartSLOduration=3.119529623 podStartE2EDuration="5.431684933s" podCreationTimestamp="2025-12-03 07:24:49 +0000 UTC" firstStartedPulling="2025-12-03 07:24:51.376321864 +0000 UTC m=+2084.173011983" lastFinishedPulling="2025-12-03 07:24:53.688477154 +0000 UTC m=+2086.485167293" observedRunningTime="2025-12-03 07:24:54.413854369 +0000 UTC m=+2087.210544488" watchObservedRunningTime="2025-12-03 07:24:54.431684933 +0000 UTC m=+2087.228375062" Dec 03 07:25:00 crc kubenswrapper[4946]: I1203 07:25:00.046514 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:25:00 crc kubenswrapper[4946]: I1203 07:25:00.047136 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:25:00 crc kubenswrapper[4946]: I1203 07:25:00.102719 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:25:00 crc kubenswrapper[4946]: I1203 07:25:00.491427 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:25:01 crc kubenswrapper[4946]: I1203 07:25:01.975896 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nx9hz"] Dec 03 07:25:02 crc kubenswrapper[4946]: I1203 07:25:02.468460 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nx9hz" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" containerName="registry-server" containerID="cri-o://654117bb1bbe9ea808dcb95f3fbf5f699dd2815df05e00fd44ab638e6da9a1e9" gracePeriod=2 Dec 03 07:25:02 crc kubenswrapper[4946]: I1203 07:25:02.985040 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zsnvl"] Dec 03 07:25:02 crc kubenswrapper[4946]: I1203 07:25:02.988957 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.004816 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsnvl"] Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.122427 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-utilities\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.122963 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gpsv\" (UniqueName: \"kubernetes.io/projected/fe337bf3-874b-467b-919e-4870f94c3c96-kube-api-access-8gpsv\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.123214 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-catalog-content\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.224301 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gpsv\" (UniqueName: \"kubernetes.io/projected/fe337bf3-874b-467b-919e-4870f94c3c96-kube-api-access-8gpsv\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.224422 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-catalog-content\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.224505 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-utilities\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.225136 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-catalog-content\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.225300 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-utilities\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.250729 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gpsv\" (UniqueName: \"kubernetes.io/projected/fe337bf3-874b-467b-919e-4870f94c3c96-kube-api-access-8gpsv\") pod \"redhat-marketplace-zsnvl\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.326141 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:03 crc kubenswrapper[4946]: I1203 07:25:03.808427 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsnvl"] Dec 03 07:25:04 crc kubenswrapper[4946]: I1203 07:25:04.485690 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsnvl" event={"ID":"fe337bf3-874b-467b-919e-4870f94c3c96","Type":"ContainerStarted","Data":"bc60eccf91c98c72efb64e6f65a1a89027b53c56a27f5700622e27c37429fea1"} Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.496059 4946 generic.go:334] "Generic (PLEG): container finished" podID="c10cd69d-808b-4196-b8c2-654f386beff6" containerID="654117bb1bbe9ea808dcb95f3fbf5f699dd2815df05e00fd44ab638e6da9a1e9" exitCode=0 Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.496389 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx9hz" event={"ID":"c10cd69d-808b-4196-b8c2-654f386beff6","Type":"ContainerDied","Data":"654117bb1bbe9ea808dcb95f3fbf5f699dd2815df05e00fd44ab638e6da9a1e9"} Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.497680 4946 generic.go:334] "Generic (PLEG): container finished" podID="fe337bf3-874b-467b-919e-4870f94c3c96" containerID="16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c" exitCode=0 Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.497705 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsnvl" event={"ID":"fe337bf3-874b-467b-919e-4870f94c3c96","Type":"ContainerDied","Data":"16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c"} Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.591391 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.764583 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-945xc\" (UniqueName: \"kubernetes.io/projected/c10cd69d-808b-4196-b8c2-654f386beff6-kube-api-access-945xc\") pod \"c10cd69d-808b-4196-b8c2-654f386beff6\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.764876 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-utilities\") pod \"c10cd69d-808b-4196-b8c2-654f386beff6\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.764916 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-catalog-content\") pod \"c10cd69d-808b-4196-b8c2-654f386beff6\" (UID: \"c10cd69d-808b-4196-b8c2-654f386beff6\") " Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.766624 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-utilities" (OuterVolumeSpecName: "utilities") pod "c10cd69d-808b-4196-b8c2-654f386beff6" (UID: "c10cd69d-808b-4196-b8c2-654f386beff6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.773527 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c10cd69d-808b-4196-b8c2-654f386beff6-kube-api-access-945xc" (OuterVolumeSpecName: "kube-api-access-945xc") pod "c10cd69d-808b-4196-b8c2-654f386beff6" (UID: "c10cd69d-808b-4196-b8c2-654f386beff6"). InnerVolumeSpecName "kube-api-access-945xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.845491 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c10cd69d-808b-4196-b8c2-654f386beff6" (UID: "c10cd69d-808b-4196-b8c2-654f386beff6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.866324 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-945xc\" (UniqueName: \"kubernetes.io/projected/c10cd69d-808b-4196-b8c2-654f386beff6-kube-api-access-945xc\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.866367 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:05 crc kubenswrapper[4946]: I1203 07:25:05.866383 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10cd69d-808b-4196-b8c2-654f386beff6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:06 crc kubenswrapper[4946]: I1203 07:25:06.511995 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx9hz" event={"ID":"c10cd69d-808b-4196-b8c2-654f386beff6","Type":"ContainerDied","Data":"009315ec4e3f97bf91463a8d387aa739551abcca2908d0a1766fbfe3e79be59c"} Dec 03 07:25:06 crc kubenswrapper[4946]: I1203 07:25:06.512105 4946 scope.go:117] "RemoveContainer" containerID="654117bb1bbe9ea808dcb95f3fbf5f699dd2815df05e00fd44ab638e6da9a1e9" Dec 03 07:25:06 crc kubenswrapper[4946]: I1203 07:25:06.513370 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx9hz" Dec 03 07:25:06 crc kubenswrapper[4946]: I1203 07:25:06.542636 4946 scope.go:117] "RemoveContainer" containerID="3a0c462461d34e4d3b7768e77aa6fe2340dc1f628adee71ec15c82dd203fa18a" Dec 03 07:25:06 crc kubenswrapper[4946]: I1203 07:25:06.579772 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nx9hz"] Dec 03 07:25:06 crc kubenswrapper[4946]: I1203 07:25:06.586923 4946 scope.go:117] "RemoveContainer" containerID="157e3638c78e60c49818aa047d144e14f3c84423af0eab47b694a7e2c06e62ea" Dec 03 07:25:06 crc kubenswrapper[4946]: I1203 07:25:06.591216 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nx9hz"] Dec 03 07:25:07 crc kubenswrapper[4946]: I1203 07:25:07.616055 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" path="/var/lib/kubelet/pods/c10cd69d-808b-4196-b8c2-654f386beff6/volumes" Dec 03 07:25:10 crc kubenswrapper[4946]: I1203 07:25:10.556415 4946 generic.go:334] "Generic (PLEG): container finished" podID="fe337bf3-874b-467b-919e-4870f94c3c96" containerID="0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee" exitCode=0 Dec 03 07:25:10 crc kubenswrapper[4946]: I1203 07:25:10.556488 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsnvl" event={"ID":"fe337bf3-874b-467b-919e-4870f94c3c96","Type":"ContainerDied","Data":"0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee"} Dec 03 07:25:11 crc kubenswrapper[4946]: I1203 07:25:11.569150 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsnvl" event={"ID":"fe337bf3-874b-467b-919e-4870f94c3c96","Type":"ContainerStarted","Data":"76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487"} Dec 03 07:25:11 crc kubenswrapper[4946]: I1203 07:25:11.596220 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zsnvl" podStartSLOduration=4.114752142 podStartE2EDuration="9.596201392s" podCreationTimestamp="2025-12-03 07:25:02 +0000 UTC" firstStartedPulling="2025-12-03 07:25:05.499129657 +0000 UTC m=+2098.295819776" lastFinishedPulling="2025-12-03 07:25:10.980578877 +0000 UTC m=+2103.777269026" observedRunningTime="2025-12-03 07:25:11.589616106 +0000 UTC m=+2104.386306225" watchObservedRunningTime="2025-12-03 07:25:11.596201392 +0000 UTC m=+2104.392891511" Dec 03 07:25:13 crc kubenswrapper[4946]: I1203 07:25:13.327024 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:13 crc kubenswrapper[4946]: I1203 07:25:13.327117 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:13 crc kubenswrapper[4946]: I1203 07:25:13.390381 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:23 crc kubenswrapper[4946]: I1203 07:25:23.375286 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:23 crc kubenswrapper[4946]: I1203 07:25:23.431439 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsnvl"] Dec 03 07:25:23 crc kubenswrapper[4946]: I1203 07:25:23.686830 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zsnvl" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" containerName="registry-server" containerID="cri-o://76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487" gracePeriod=2 Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.161984 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.212335 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-catalog-content\") pod \"fe337bf3-874b-467b-919e-4870f94c3c96\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.212388 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-utilities\") pod \"fe337bf3-874b-467b-919e-4870f94c3c96\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.212500 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gpsv\" (UniqueName: \"kubernetes.io/projected/fe337bf3-874b-467b-919e-4870f94c3c96-kube-api-access-8gpsv\") pod \"fe337bf3-874b-467b-919e-4870f94c3c96\" (UID: \"fe337bf3-874b-467b-919e-4870f94c3c96\") " Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.213692 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-utilities" (OuterVolumeSpecName: "utilities") pod "fe337bf3-874b-467b-919e-4870f94c3c96" (UID: "fe337bf3-874b-467b-919e-4870f94c3c96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.217320 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe337bf3-874b-467b-919e-4870f94c3c96-kube-api-access-8gpsv" (OuterVolumeSpecName: "kube-api-access-8gpsv") pod "fe337bf3-874b-467b-919e-4870f94c3c96" (UID: "fe337bf3-874b-467b-919e-4870f94c3c96"). InnerVolumeSpecName "kube-api-access-8gpsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.248666 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe337bf3-874b-467b-919e-4870f94c3c96" (UID: "fe337bf3-874b-467b-919e-4870f94c3c96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.314078 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.314116 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe337bf3-874b-467b-919e-4870f94c3c96-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.314132 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gpsv\" (UniqueName: \"kubernetes.io/projected/fe337bf3-874b-467b-919e-4870f94c3c96-kube-api-access-8gpsv\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.699054 4946 generic.go:334] "Generic (PLEG): container finished" podID="fe337bf3-874b-467b-919e-4870f94c3c96" containerID="76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487" exitCode=0 Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.699158 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsnvl" event={"ID":"fe337bf3-874b-467b-919e-4870f94c3c96","Type":"ContainerDied","Data":"76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487"} Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.699181 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zsnvl" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.699207 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsnvl" event={"ID":"fe337bf3-874b-467b-919e-4870f94c3c96","Type":"ContainerDied","Data":"bc60eccf91c98c72efb64e6f65a1a89027b53c56a27f5700622e27c37429fea1"} Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.699243 4946 scope.go:117] "RemoveContainer" containerID="76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.754602 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsnvl"] Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.757837 4946 scope.go:117] "RemoveContainer" containerID="0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.768791 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsnvl"] Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.795644 4946 scope.go:117] "RemoveContainer" containerID="16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.826047 4946 scope.go:117] "RemoveContainer" containerID="76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487" Dec 03 07:25:24 crc kubenswrapper[4946]: E1203 07:25:24.827942 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487\": container with ID starting with 76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487 not found: ID does not exist" containerID="76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.827989 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487"} err="failed to get container status \"76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487\": rpc error: code = NotFound desc = could not find container \"76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487\": container with ID starting with 76e712a9d45a595a139f63e7bfdae9b6c7d4f409cc6c89c90f800d3e12819487 not found: ID does not exist" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.828015 4946 scope.go:117] "RemoveContainer" containerID="0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee" Dec 03 07:25:24 crc kubenswrapper[4946]: E1203 07:25:24.828682 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee\": container with ID starting with 0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee not found: ID does not exist" containerID="0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.828714 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee"} err="failed to get container status \"0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee\": rpc error: code = NotFound desc = could not find container \"0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee\": container with ID starting with 0e9fc590c4d0a1e731341074396c43ce70c5ea8c508e80d98b2367535d4575ee not found: ID does not exist" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.828731 4946 scope.go:117] "RemoveContainer" containerID="16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c" Dec 03 07:25:24 crc kubenswrapper[4946]: E1203 07:25:24.829536 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c\": container with ID starting with 16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c not found: ID does not exist" containerID="16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c" Dec 03 07:25:24 crc kubenswrapper[4946]: I1203 07:25:24.829595 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c"} err="failed to get container status \"16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c\": rpc error: code = NotFound desc = could not find container \"16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c\": container with ID starting with 16bf993516a0fc1efa8bff0d82b6f0f474dd3ef5b29bb44edadd99736f58267c not found: ID does not exist" Dec 03 07:25:25 crc kubenswrapper[4946]: I1203 07:25:25.610225 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" path="/var/lib/kubelet/pods/fe337bf3-874b-467b-919e-4870f94c3c96/volumes" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.772205 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l5ltb"] Dec 03 07:25:56 crc kubenswrapper[4946]: E1203 07:25:56.773031 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" containerName="extract-content" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.773045 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" containerName="extract-content" Dec 03 07:25:56 crc kubenswrapper[4946]: E1203 07:25:56.773071 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" containerName="registry-server" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.773078 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" containerName="registry-server" Dec 03 07:25:56 crc kubenswrapper[4946]: E1203 07:25:56.773097 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" containerName="extract-utilities" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.773105 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" containerName="extract-utilities" Dec 03 07:25:56 crc kubenswrapper[4946]: E1203 07:25:56.773118 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" containerName="extract-content" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.773125 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" containerName="extract-content" Dec 03 07:25:56 crc kubenswrapper[4946]: E1203 07:25:56.773142 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" containerName="registry-server" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.773150 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" containerName="registry-server" Dec 03 07:25:56 crc kubenswrapper[4946]: E1203 07:25:56.773162 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" containerName="extract-utilities" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.773169 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" containerName="extract-utilities" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.773337 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe337bf3-874b-467b-919e-4870f94c3c96" containerName="registry-server" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.773355 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c10cd69d-808b-4196-b8c2-654f386beff6" containerName="registry-server" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.774493 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.800624 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5ltb"] Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.845637 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-catalog-content\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.846018 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvrlz\" (UniqueName: \"kubernetes.io/projected/208dd19a-e63b-4f3a-8f8a-f307235693a8-kube-api-access-nvrlz\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.846269 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-utilities\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.947881 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-utilities\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.947990 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-catalog-content\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.948036 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvrlz\" (UniqueName: \"kubernetes.io/projected/208dd19a-e63b-4f3a-8f8a-f307235693a8-kube-api-access-nvrlz\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.948544 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-utilities\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.948613 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-catalog-content\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:56 crc kubenswrapper[4946]: I1203 07:25:56.972935 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvrlz\" (UniqueName: \"kubernetes.io/projected/208dd19a-e63b-4f3a-8f8a-f307235693a8-kube-api-access-nvrlz\") pod \"redhat-operators-l5ltb\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:57 crc kubenswrapper[4946]: I1203 07:25:57.111957 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:25:57 crc kubenswrapper[4946]: I1203 07:25:57.522641 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5ltb"] Dec 03 07:25:58 crc kubenswrapper[4946]: I1203 07:25:58.035782 4946 generic.go:334] "Generic (PLEG): container finished" podID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerID="39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0" exitCode=0 Dec 03 07:25:58 crc kubenswrapper[4946]: I1203 07:25:58.035864 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5ltb" event={"ID":"208dd19a-e63b-4f3a-8f8a-f307235693a8","Type":"ContainerDied","Data":"39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0"} Dec 03 07:25:58 crc kubenswrapper[4946]: I1203 07:25:58.036112 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5ltb" event={"ID":"208dd19a-e63b-4f3a-8f8a-f307235693a8","Type":"ContainerStarted","Data":"e87dffb56940734120b59121d8e2282d423394f9a709fa8ac63563bfd1606d3a"} Dec 03 07:26:00 crc kubenswrapper[4946]: I1203 07:26:00.050344 4946 generic.go:334] "Generic (PLEG): container finished" podID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerID="91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b" exitCode=0 Dec 03 07:26:00 crc kubenswrapper[4946]: I1203 07:26:00.050432 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5ltb" event={"ID":"208dd19a-e63b-4f3a-8f8a-f307235693a8","Type":"ContainerDied","Data":"91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b"} Dec 03 07:26:01 crc kubenswrapper[4946]: I1203 07:26:01.064251 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5ltb" event={"ID":"208dd19a-e63b-4f3a-8f8a-f307235693a8","Type":"ContainerStarted","Data":"00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72"} Dec 03 07:26:01 crc kubenswrapper[4946]: I1203 07:26:01.091067 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l5ltb" podStartSLOduration=2.57153575 podStartE2EDuration="5.091031505s" podCreationTimestamp="2025-12-03 07:25:56 +0000 UTC" firstStartedPulling="2025-12-03 07:25:58.03690598 +0000 UTC m=+2150.833596089" lastFinishedPulling="2025-12-03 07:26:00.556401705 +0000 UTC m=+2153.353091844" observedRunningTime="2025-12-03 07:26:01.089034342 +0000 UTC m=+2153.885724551" watchObservedRunningTime="2025-12-03 07:26:01.091031505 +0000 UTC m=+2153.887721674" Dec 03 07:26:07 crc kubenswrapper[4946]: I1203 07:26:07.112931 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:26:07 crc kubenswrapper[4946]: I1203 07:26:07.113686 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:26:07 crc kubenswrapper[4946]: I1203 07:26:07.192251 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:26:08 crc kubenswrapper[4946]: I1203 07:26:08.197054 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:26:10 crc kubenswrapper[4946]: I1203 07:26:10.683040 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5ltb"] Dec 03 07:26:10 crc kubenswrapper[4946]: I1203 07:26:10.683874 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l5ltb" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerName="registry-server" containerID="cri-o://00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72" gracePeriod=2 Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.064070 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.082638 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvrlz\" (UniqueName: \"kubernetes.io/projected/208dd19a-e63b-4f3a-8f8a-f307235693a8-kube-api-access-nvrlz\") pod \"208dd19a-e63b-4f3a-8f8a-f307235693a8\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.082691 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-catalog-content\") pod \"208dd19a-e63b-4f3a-8f8a-f307235693a8\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.082857 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-utilities\") pod \"208dd19a-e63b-4f3a-8f8a-f307235693a8\" (UID: \"208dd19a-e63b-4f3a-8f8a-f307235693a8\") " Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.083693 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-utilities" (OuterVolumeSpecName: "utilities") pod "208dd19a-e63b-4f3a-8f8a-f307235693a8" (UID: "208dd19a-e63b-4f3a-8f8a-f307235693a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.100820 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/208dd19a-e63b-4f3a-8f8a-f307235693a8-kube-api-access-nvrlz" (OuterVolumeSpecName: "kube-api-access-nvrlz") pod "208dd19a-e63b-4f3a-8f8a-f307235693a8" (UID: "208dd19a-e63b-4f3a-8f8a-f307235693a8"). InnerVolumeSpecName "kube-api-access-nvrlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.177065 4946 generic.go:334] "Generic (PLEG): container finished" podID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerID="00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72" exitCode=0 Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.177113 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5ltb" event={"ID":"208dd19a-e63b-4f3a-8f8a-f307235693a8","Type":"ContainerDied","Data":"00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72"} Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.177139 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5ltb" event={"ID":"208dd19a-e63b-4f3a-8f8a-f307235693a8","Type":"ContainerDied","Data":"e87dffb56940734120b59121d8e2282d423394f9a709fa8ac63563bfd1606d3a"} Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.177161 4946 scope.go:117] "RemoveContainer" containerID="00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.177179 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5ltb" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.184562 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvrlz\" (UniqueName: \"kubernetes.io/projected/208dd19a-e63b-4f3a-8f8a-f307235693a8-kube-api-access-nvrlz\") on node \"crc\" DevicePath \"\"" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.184609 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.192970 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "208dd19a-e63b-4f3a-8f8a-f307235693a8" (UID: "208dd19a-e63b-4f3a-8f8a-f307235693a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.201649 4946 scope.go:117] "RemoveContainer" containerID="91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.225301 4946 scope.go:117] "RemoveContainer" containerID="39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.247790 4946 scope.go:117] "RemoveContainer" containerID="00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72" Dec 03 07:26:13 crc kubenswrapper[4946]: E1203 07:26:13.248307 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72\": container with ID starting with 00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72 not found: ID does not exist" containerID="00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.248380 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72"} err="failed to get container status \"00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72\": rpc error: code = NotFound desc = could not find container \"00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72\": container with ID starting with 00d258798a6324a325e081f591c499863f801438a15e3f132b95e3932928fb72 not found: ID does not exist" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.248413 4946 scope.go:117] "RemoveContainer" containerID="91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b" Dec 03 07:26:13 crc kubenswrapper[4946]: E1203 07:26:13.248881 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b\": container with ID starting with 91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b not found: ID does not exist" containerID="91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.248951 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b"} err="failed to get container status \"91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b\": rpc error: code = NotFound desc = could not find container \"91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b\": container with ID starting with 91d405da37371f2c2bd8d5cd3d6ee03a0f76620a2b494deaa46fe108fd2ccb5b not found: ID does not exist" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.248996 4946 scope.go:117] "RemoveContainer" containerID="39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0" Dec 03 07:26:13 crc kubenswrapper[4946]: E1203 07:26:13.249390 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0\": container with ID starting with 39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0 not found: ID does not exist" containerID="39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.249453 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0"} err="failed to get container status \"39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0\": rpc error: code = NotFound desc = could not find container \"39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0\": container with ID starting with 39fde484866e2048b9108c632862cb17faf23954afea98daf067465f5980a8f0 not found: ID does not exist" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.286077 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/208dd19a-e63b-4f3a-8f8a-f307235693a8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.514299 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5ltb"] Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.520272 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l5ltb"] Dec 03 07:26:13 crc kubenswrapper[4946]: I1203 07:26:13.603851 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" path="/var/lib/kubelet/pods/208dd19a-e63b-4f3a-8f8a-f307235693a8/volumes" Dec 03 07:26:53 crc kubenswrapper[4946]: I1203 07:26:53.039788 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:26:53 crc kubenswrapper[4946]: I1203 07:26:53.040483 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:27:23 crc kubenswrapper[4946]: I1203 07:27:23.040073 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:27:23 crc kubenswrapper[4946]: I1203 07:27:23.040904 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:27:53 crc kubenswrapper[4946]: I1203 07:27:53.039197 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:27:53 crc kubenswrapper[4946]: I1203 07:27:53.039930 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:27:53 crc kubenswrapper[4946]: I1203 07:27:53.039992 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:27:53 crc kubenswrapper[4946]: I1203 07:27:53.040925 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:27:53 crc kubenswrapper[4946]: I1203 07:27:53.041009 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" gracePeriod=600 Dec 03 07:27:53 crc kubenswrapper[4946]: E1203 07:27:53.684384 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:27:54 crc kubenswrapper[4946]: I1203 07:27:54.135867 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" exitCode=0 Dec 03 07:27:54 crc kubenswrapper[4946]: I1203 07:27:54.135903 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad"} Dec 03 07:27:54 crc kubenswrapper[4946]: I1203 07:27:54.135969 4946 scope.go:117] "RemoveContainer" containerID="f63f6469fc21ed7b131500e8679e2fed6eb0e75d27f7ea70e25254efa4de3f73" Dec 03 07:27:54 crc kubenswrapper[4946]: I1203 07:27:54.138313 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:27:54 crc kubenswrapper[4946]: E1203 07:27:54.138646 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:28:08 crc kubenswrapper[4946]: I1203 07:28:08.592536 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:28:08 crc kubenswrapper[4946]: E1203 07:28:08.593531 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.309824 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6gjb9"] Dec 03 07:28:13 crc kubenswrapper[4946]: E1203 07:28:13.310975 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerName="extract-utilities" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.311014 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerName="extract-utilities" Dec 03 07:28:13 crc kubenswrapper[4946]: E1203 07:28:13.311026 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerName="extract-content" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.311034 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerName="extract-content" Dec 03 07:28:13 crc kubenswrapper[4946]: E1203 07:28:13.311052 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerName="registry-server" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.311060 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerName="registry-server" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.311294 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="208dd19a-e63b-4f3a-8f8a-f307235693a8" containerName="registry-server" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.313107 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.334307 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6gjb9"] Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.335053 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-utilities\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.335229 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd5rc\" (UniqueName: \"kubernetes.io/projected/aef94e6e-05bc-4700-808e-49a38d6af87c-kube-api-access-kd5rc\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.335316 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-catalog-content\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.436409 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-utilities\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.436482 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd5rc\" (UniqueName: \"kubernetes.io/projected/aef94e6e-05bc-4700-808e-49a38d6af87c-kube-api-access-kd5rc\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.436504 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-catalog-content\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.437049 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-catalog-content\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.437525 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-utilities\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.476359 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd5rc\" (UniqueName: \"kubernetes.io/projected/aef94e6e-05bc-4700-808e-49a38d6af87c-kube-api-access-kd5rc\") pod \"community-operators-6gjb9\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.645208 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:13 crc kubenswrapper[4946]: I1203 07:28:13.925537 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6gjb9"] Dec 03 07:28:14 crc kubenswrapper[4946]: I1203 07:28:14.331223 4946 generic.go:334] "Generic (PLEG): container finished" podID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerID="2df9b5cab48cdb2206ed4c50801004502a7f31518a4de2d674338bc538007c84" exitCode=0 Dec 03 07:28:14 crc kubenswrapper[4946]: I1203 07:28:14.331303 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gjb9" event={"ID":"aef94e6e-05bc-4700-808e-49a38d6af87c","Type":"ContainerDied","Data":"2df9b5cab48cdb2206ed4c50801004502a7f31518a4de2d674338bc538007c84"} Dec 03 07:28:14 crc kubenswrapper[4946]: I1203 07:28:14.331348 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gjb9" event={"ID":"aef94e6e-05bc-4700-808e-49a38d6af87c","Type":"ContainerStarted","Data":"f5946dae4fcd51d072bc5fb74f0deb97830c572587342c96590d484002fc597d"} Dec 03 07:28:15 crc kubenswrapper[4946]: I1203 07:28:15.341649 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gjb9" event={"ID":"aef94e6e-05bc-4700-808e-49a38d6af87c","Type":"ContainerStarted","Data":"bd1a694bd510781e6a3690003410e05358ab6c7eef1a8bc7cd9eae83b75168d4"} Dec 03 07:28:16 crc kubenswrapper[4946]: I1203 07:28:16.354418 4946 generic.go:334] "Generic (PLEG): container finished" podID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerID="bd1a694bd510781e6a3690003410e05358ab6c7eef1a8bc7cd9eae83b75168d4" exitCode=0 Dec 03 07:28:16 crc kubenswrapper[4946]: I1203 07:28:16.354491 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gjb9" event={"ID":"aef94e6e-05bc-4700-808e-49a38d6af87c","Type":"ContainerDied","Data":"bd1a694bd510781e6a3690003410e05358ab6c7eef1a8bc7cd9eae83b75168d4"} Dec 03 07:28:17 crc kubenswrapper[4946]: I1203 07:28:17.366087 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gjb9" event={"ID":"aef94e6e-05bc-4700-808e-49a38d6af87c","Type":"ContainerStarted","Data":"8ff73d0f868838dce183831ad2d1046a78c89f3ff68d9f74ea1bddba1dc632b8"} Dec 03 07:28:17 crc kubenswrapper[4946]: I1203 07:28:17.409669 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6gjb9" podStartSLOduration=1.95003568 podStartE2EDuration="4.409648715s" podCreationTimestamp="2025-12-03 07:28:13 +0000 UTC" firstStartedPulling="2025-12-03 07:28:14.33647932 +0000 UTC m=+2287.133169439" lastFinishedPulling="2025-12-03 07:28:16.796092365 +0000 UTC m=+2289.592782474" observedRunningTime="2025-12-03 07:28:17.383426115 +0000 UTC m=+2290.180116214" watchObservedRunningTime="2025-12-03 07:28:17.409648715 +0000 UTC m=+2290.206338824" Dec 03 07:28:19 crc kubenswrapper[4946]: I1203 07:28:19.593051 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:28:19 crc kubenswrapper[4946]: E1203 07:28:19.593465 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:28:23 crc kubenswrapper[4946]: I1203 07:28:23.645321 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:23 crc kubenswrapper[4946]: I1203 07:28:23.645703 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:23 crc kubenswrapper[4946]: I1203 07:28:23.734199 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:24 crc kubenswrapper[4946]: I1203 07:28:24.721529 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:24 crc kubenswrapper[4946]: I1203 07:28:24.767793 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6gjb9"] Dec 03 07:28:26 crc kubenswrapper[4946]: I1203 07:28:26.668391 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6gjb9" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerName="registry-server" containerID="cri-o://8ff73d0f868838dce183831ad2d1046a78c89f3ff68d9f74ea1bddba1dc632b8" gracePeriod=2 Dec 03 07:28:27 crc kubenswrapper[4946]: I1203 07:28:27.693692 4946 generic.go:334] "Generic (PLEG): container finished" podID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerID="8ff73d0f868838dce183831ad2d1046a78c89f3ff68d9f74ea1bddba1dc632b8" exitCode=0 Dec 03 07:28:27 crc kubenswrapper[4946]: I1203 07:28:27.693908 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gjb9" event={"ID":"aef94e6e-05bc-4700-808e-49a38d6af87c","Type":"ContainerDied","Data":"8ff73d0f868838dce183831ad2d1046a78c89f3ff68d9f74ea1bddba1dc632b8"} Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.286163 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.437792 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-utilities\") pod \"aef94e6e-05bc-4700-808e-49a38d6af87c\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.438191 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kd5rc\" (UniqueName: \"kubernetes.io/projected/aef94e6e-05bc-4700-808e-49a38d6af87c-kube-api-access-kd5rc\") pod \"aef94e6e-05bc-4700-808e-49a38d6af87c\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.439635 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-catalog-content\") pod \"aef94e6e-05bc-4700-808e-49a38d6af87c\" (UID: \"aef94e6e-05bc-4700-808e-49a38d6af87c\") " Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.440840 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-utilities" (OuterVolumeSpecName: "utilities") pod "aef94e6e-05bc-4700-808e-49a38d6af87c" (UID: "aef94e6e-05bc-4700-808e-49a38d6af87c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.446451 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef94e6e-05bc-4700-808e-49a38d6af87c-kube-api-access-kd5rc" (OuterVolumeSpecName: "kube-api-access-kd5rc") pod "aef94e6e-05bc-4700-808e-49a38d6af87c" (UID: "aef94e6e-05bc-4700-808e-49a38d6af87c"). InnerVolumeSpecName "kube-api-access-kd5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.511081 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aef94e6e-05bc-4700-808e-49a38d6af87c" (UID: "aef94e6e-05bc-4700-808e-49a38d6af87c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.541845 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.541888 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef94e6e-05bc-4700-808e-49a38d6af87c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.541902 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kd5rc\" (UniqueName: \"kubernetes.io/projected/aef94e6e-05bc-4700-808e-49a38d6af87c-kube-api-access-kd5rc\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.706570 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6gjb9" event={"ID":"aef94e6e-05bc-4700-808e-49a38d6af87c","Type":"ContainerDied","Data":"f5946dae4fcd51d072bc5fb74f0deb97830c572587342c96590d484002fc597d"} Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.706652 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6gjb9" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.706671 4946 scope.go:117] "RemoveContainer" containerID="8ff73d0f868838dce183831ad2d1046a78c89f3ff68d9f74ea1bddba1dc632b8" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.744164 4946 scope.go:117] "RemoveContainer" containerID="bd1a694bd510781e6a3690003410e05358ab6c7eef1a8bc7cd9eae83b75168d4" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.779636 4946 scope.go:117] "RemoveContainer" containerID="2df9b5cab48cdb2206ed4c50801004502a7f31518a4de2d674338bc538007c84" Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.788179 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6gjb9"] Dec 03 07:28:28 crc kubenswrapper[4946]: I1203 07:28:28.803113 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6gjb9"] Dec 03 07:28:29 crc kubenswrapper[4946]: I1203 07:28:29.605435 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" path="/var/lib/kubelet/pods/aef94e6e-05bc-4700-808e-49a38d6af87c/volumes" Dec 03 07:28:32 crc kubenswrapper[4946]: I1203 07:28:32.593567 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:28:32 crc kubenswrapper[4946]: E1203 07:28:32.594371 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:28:44 crc kubenswrapper[4946]: I1203 07:28:44.592918 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:28:44 crc kubenswrapper[4946]: E1203 07:28:44.593574 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:28:58 crc kubenswrapper[4946]: I1203 07:28:58.592946 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:28:58 crc kubenswrapper[4946]: E1203 07:28:58.593971 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:29:09 crc kubenswrapper[4946]: I1203 07:29:09.593397 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:29:09 crc kubenswrapper[4946]: E1203 07:29:09.594044 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:29:20 crc kubenswrapper[4946]: I1203 07:29:20.592629 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:29:20 crc kubenswrapper[4946]: E1203 07:29:20.593281 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:29:34 crc kubenswrapper[4946]: I1203 07:29:34.593612 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:29:34 crc kubenswrapper[4946]: E1203 07:29:34.594612 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:29:49 crc kubenswrapper[4946]: I1203 07:29:49.592827 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:29:49 crc kubenswrapper[4946]: E1203 07:29:49.594158 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.164073 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf"] Dec 03 07:30:00 crc kubenswrapper[4946]: E1203 07:30:00.165424 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerName="extract-content" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.165460 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerName="extract-content" Dec 03 07:30:00 crc kubenswrapper[4946]: E1203 07:30:00.165520 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerName="registry-server" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.165535 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerName="registry-server" Dec 03 07:30:00 crc kubenswrapper[4946]: E1203 07:30:00.165553 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerName="extract-utilities" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.165568 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerName="extract-utilities" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.165946 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef94e6e-05bc-4700-808e-49a38d6af87c" containerName="registry-server" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.167270 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.170608 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.170679 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.176439 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf"] Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.277351 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0069d43-de07-40a9-84aa-11d303fb01b0-config-volume\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.277432 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0069d43-de07-40a9-84aa-11d303fb01b0-secret-volume\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.277455 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rptnr\" (UniqueName: \"kubernetes.io/projected/d0069d43-de07-40a9-84aa-11d303fb01b0-kube-api-access-rptnr\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.379265 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0069d43-de07-40a9-84aa-11d303fb01b0-secret-volume\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.379323 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rptnr\" (UniqueName: \"kubernetes.io/projected/d0069d43-de07-40a9-84aa-11d303fb01b0-kube-api-access-rptnr\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.379412 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0069d43-de07-40a9-84aa-11d303fb01b0-config-volume\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.380610 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0069d43-de07-40a9-84aa-11d303fb01b0-config-volume\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.387857 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0069d43-de07-40a9-84aa-11d303fb01b0-secret-volume\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.398538 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rptnr\" (UniqueName: \"kubernetes.io/projected/d0069d43-de07-40a9-84aa-11d303fb01b0-kube-api-access-rptnr\") pod \"collect-profiles-29412450-wxddf\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.497624 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:00 crc kubenswrapper[4946]: I1203 07:30:00.754348 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf"] Dec 03 07:30:01 crc kubenswrapper[4946]: E1203 07:30:01.280090 4946 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0069d43_de07_40a9_84aa_11d303fb01b0.slice/crio-conmon-02f75138b6c78b38777921bacfcc64dd4078032bd386a56d9d7894bd8f2e10c3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0069d43_de07_40a9_84aa_11d303fb01b0.slice/crio-02f75138b6c78b38777921bacfcc64dd4078032bd386a56d9d7894bd8f2e10c3.scope\": RecentStats: unable to find data in memory cache]" Dec 03 07:30:01 crc kubenswrapper[4946]: I1203 07:30:01.592678 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:30:01 crc kubenswrapper[4946]: E1203 07:30:01.593443 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:30:01 crc kubenswrapper[4946]: I1203 07:30:01.610634 4946 generic.go:334] "Generic (PLEG): container finished" podID="d0069d43-de07-40a9-84aa-11d303fb01b0" containerID="02f75138b6c78b38777921bacfcc64dd4078032bd386a56d9d7894bd8f2e10c3" exitCode=0 Dec 03 07:30:01 crc kubenswrapper[4946]: I1203 07:30:01.613245 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" event={"ID":"d0069d43-de07-40a9-84aa-11d303fb01b0","Type":"ContainerDied","Data":"02f75138b6c78b38777921bacfcc64dd4078032bd386a56d9d7894bd8f2e10c3"} Dec 03 07:30:01 crc kubenswrapper[4946]: I1203 07:30:01.613313 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" event={"ID":"d0069d43-de07-40a9-84aa-11d303fb01b0","Type":"ContainerStarted","Data":"94c1e782749902bb2205c9608107aa783f26a1e8c3ecd328e4ca5b0e3c94e29d"} Dec 03 07:30:02 crc kubenswrapper[4946]: I1203 07:30:02.943053 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.024022 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rptnr\" (UniqueName: \"kubernetes.io/projected/d0069d43-de07-40a9-84aa-11d303fb01b0-kube-api-access-rptnr\") pod \"d0069d43-de07-40a9-84aa-11d303fb01b0\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.024130 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0069d43-de07-40a9-84aa-11d303fb01b0-secret-volume\") pod \"d0069d43-de07-40a9-84aa-11d303fb01b0\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.024186 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0069d43-de07-40a9-84aa-11d303fb01b0-config-volume\") pod \"d0069d43-de07-40a9-84aa-11d303fb01b0\" (UID: \"d0069d43-de07-40a9-84aa-11d303fb01b0\") " Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.025204 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0069d43-de07-40a9-84aa-11d303fb01b0-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0069d43-de07-40a9-84aa-11d303fb01b0" (UID: "d0069d43-de07-40a9-84aa-11d303fb01b0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.032366 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0069d43-de07-40a9-84aa-11d303fb01b0-kube-api-access-rptnr" (OuterVolumeSpecName: "kube-api-access-rptnr") pod "d0069d43-de07-40a9-84aa-11d303fb01b0" (UID: "d0069d43-de07-40a9-84aa-11d303fb01b0"). InnerVolumeSpecName "kube-api-access-rptnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.032365 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0069d43-de07-40a9-84aa-11d303fb01b0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0069d43-de07-40a9-84aa-11d303fb01b0" (UID: "d0069d43-de07-40a9-84aa-11d303fb01b0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.126278 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0069d43-de07-40a9-84aa-11d303fb01b0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.126340 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0069d43-de07-40a9-84aa-11d303fb01b0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.126361 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rptnr\" (UniqueName: \"kubernetes.io/projected/d0069d43-de07-40a9-84aa-11d303fb01b0-kube-api-access-rptnr\") on node \"crc\" DevicePath \"\"" Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.629305 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" event={"ID":"d0069d43-de07-40a9-84aa-11d303fb01b0","Type":"ContainerDied","Data":"94c1e782749902bb2205c9608107aa783f26a1e8c3ecd328e4ca5b0e3c94e29d"} Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.629619 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94c1e782749902bb2205c9608107aa783f26a1e8c3ecd328e4ca5b0e3c94e29d" Dec 03 07:30:03 crc kubenswrapper[4946]: I1203 07:30:03.629343 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf" Dec 03 07:30:04 crc kubenswrapper[4946]: I1203 07:30:04.043852 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd"] Dec 03 07:30:04 crc kubenswrapper[4946]: I1203 07:30:04.051156 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-8gzbd"] Dec 03 07:30:05 crc kubenswrapper[4946]: I1203 07:30:05.609558 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c58dec3d-f028-43b2-8e71-1954da44b01c" path="/var/lib/kubelet/pods/c58dec3d-f028-43b2-8e71-1954da44b01c/volumes" Dec 03 07:30:12 crc kubenswrapper[4946]: I1203 07:30:12.954784 4946 scope.go:117] "RemoveContainer" containerID="94f485723b01d8266eef0172918fa3052a8d3ef7cfb2c59b2025e56ad061883d" Dec 03 07:30:16 crc kubenswrapper[4946]: I1203 07:30:16.594115 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:30:16 crc kubenswrapper[4946]: E1203 07:30:16.594873 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:30:29 crc kubenswrapper[4946]: I1203 07:30:29.593261 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:30:29 crc kubenswrapper[4946]: E1203 07:30:29.594462 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:30:44 crc kubenswrapper[4946]: I1203 07:30:44.593185 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:30:44 crc kubenswrapper[4946]: E1203 07:30:44.594327 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:30:55 crc kubenswrapper[4946]: I1203 07:30:55.593027 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:30:55 crc kubenswrapper[4946]: E1203 07:30:55.593714 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:31:08 crc kubenswrapper[4946]: I1203 07:31:08.594107 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:31:08 crc kubenswrapper[4946]: E1203 07:31:08.595097 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:31:23 crc kubenswrapper[4946]: I1203 07:31:23.593367 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:31:23 crc kubenswrapper[4946]: E1203 07:31:23.594206 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:31:34 crc kubenswrapper[4946]: I1203 07:31:34.592975 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:31:34 crc kubenswrapper[4946]: E1203 07:31:34.593682 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:31:45 crc kubenswrapper[4946]: I1203 07:31:45.593176 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:31:45 crc kubenswrapper[4946]: E1203 07:31:45.593780 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:31:57 crc kubenswrapper[4946]: I1203 07:31:57.603780 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:31:57 crc kubenswrapper[4946]: E1203 07:31:57.605000 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:32:09 crc kubenswrapper[4946]: I1203 07:32:09.593789 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:32:09 crc kubenswrapper[4946]: E1203 07:32:09.594938 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:32:21 crc kubenswrapper[4946]: I1203 07:32:21.592804 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:32:21 crc kubenswrapper[4946]: E1203 07:32:21.593925 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:32:34 crc kubenswrapper[4946]: I1203 07:32:34.594725 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:32:34 crc kubenswrapper[4946]: E1203 07:32:34.595760 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:32:47 crc kubenswrapper[4946]: I1203 07:32:47.603318 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:32:47 crc kubenswrapper[4946]: E1203 07:32:47.603977 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:33:01 crc kubenswrapper[4946]: I1203 07:33:01.593137 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:33:02 crc kubenswrapper[4946]: I1203 07:33:02.615018 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"8d78922732bd24992d7ce5ed423db2dc2eddfc0231ce52f4145d041f2ec118b6"} Dec 03 07:35:03 crc kubenswrapper[4946]: I1203 07:35:03.792727 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z5nr8"] Dec 03 07:35:03 crc kubenswrapper[4946]: E1203 07:35:03.793714 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0069d43-de07-40a9-84aa-11d303fb01b0" containerName="collect-profiles" Dec 03 07:35:03 crc kubenswrapper[4946]: I1203 07:35:03.793756 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0069d43-de07-40a9-84aa-11d303fb01b0" containerName="collect-profiles" Dec 03 07:35:03 crc kubenswrapper[4946]: I1203 07:35:03.793985 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0069d43-de07-40a9-84aa-11d303fb01b0" containerName="collect-profiles" Dec 03 07:35:03 crc kubenswrapper[4946]: I1203 07:35:03.797096 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:03 crc kubenswrapper[4946]: I1203 07:35:03.817641 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5nr8"] Dec 03 07:35:03 crc kubenswrapper[4946]: I1203 07:35:03.982998 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-catalog-content\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:03 crc kubenswrapper[4946]: I1203 07:35:03.983260 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqw6p\" (UniqueName: \"kubernetes.io/projected/0cf1d6a8-6356-4592-ae19-884f4b15bed6-kube-api-access-bqw6p\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:03 crc kubenswrapper[4946]: I1203 07:35:03.983353 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-utilities\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.084347 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqw6p\" (UniqueName: \"kubernetes.io/projected/0cf1d6a8-6356-4592-ae19-884f4b15bed6-kube-api-access-bqw6p\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.084410 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-utilities\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.084506 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-catalog-content\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.085097 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-catalog-content\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.085313 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-utilities\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.112170 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqw6p\" (UniqueName: \"kubernetes.io/projected/0cf1d6a8-6356-4592-ae19-884f4b15bed6-kube-api-access-bqw6p\") pod \"redhat-marketplace-z5nr8\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.154344 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.622164 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5nr8"] Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.853549 4946 generic.go:334] "Generic (PLEG): container finished" podID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerID="0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750" exitCode=0 Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.853794 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5nr8" event={"ID":"0cf1d6a8-6356-4592-ae19-884f4b15bed6","Type":"ContainerDied","Data":"0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750"} Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.853867 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5nr8" event={"ID":"0cf1d6a8-6356-4592-ae19-884f4b15bed6","Type":"ContainerStarted","Data":"53b25b5293ad6cb98b2ed72825a7e1d6fa1d81e5d7bacf40363e088c83e94b47"} Dec 03 07:35:04 crc kubenswrapper[4946]: I1203 07:35:04.855200 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:35:06 crc kubenswrapper[4946]: I1203 07:35:06.875203 4946 generic.go:334] "Generic (PLEG): container finished" podID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerID="afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef" exitCode=0 Dec 03 07:35:06 crc kubenswrapper[4946]: I1203 07:35:06.875338 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5nr8" event={"ID":"0cf1d6a8-6356-4592-ae19-884f4b15bed6","Type":"ContainerDied","Data":"afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef"} Dec 03 07:35:07 crc kubenswrapper[4946]: I1203 07:35:07.886966 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5nr8" event={"ID":"0cf1d6a8-6356-4592-ae19-884f4b15bed6","Type":"ContainerStarted","Data":"b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5"} Dec 03 07:35:07 crc kubenswrapper[4946]: I1203 07:35:07.907116 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z5nr8" podStartSLOduration=2.485250857 podStartE2EDuration="4.907094277s" podCreationTimestamp="2025-12-03 07:35:03 +0000 UTC" firstStartedPulling="2025-12-03 07:35:04.855009966 +0000 UTC m=+2697.651700075" lastFinishedPulling="2025-12-03 07:35:07.276853386 +0000 UTC m=+2700.073543495" observedRunningTime="2025-12-03 07:35:07.903690036 +0000 UTC m=+2700.700380155" watchObservedRunningTime="2025-12-03 07:35:07.907094277 +0000 UTC m=+2700.703784386" Dec 03 07:35:14 crc kubenswrapper[4946]: I1203 07:35:14.155313 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:14 crc kubenswrapper[4946]: I1203 07:35:14.155903 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:14 crc kubenswrapper[4946]: I1203 07:35:14.220834 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:15 crc kubenswrapper[4946]: I1203 07:35:15.010936 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:15 crc kubenswrapper[4946]: I1203 07:35:15.077237 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5nr8"] Dec 03 07:35:16 crc kubenswrapper[4946]: I1203 07:35:16.957664 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z5nr8" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerName="registry-server" containerID="cri-o://b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5" gracePeriod=2 Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.843496 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.887105 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-utilities\") pod \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.887279 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqw6p\" (UniqueName: \"kubernetes.io/projected/0cf1d6a8-6356-4592-ae19-884f4b15bed6-kube-api-access-bqw6p\") pod \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.887343 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-catalog-content\") pod \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\" (UID: \"0cf1d6a8-6356-4592-ae19-884f4b15bed6\") " Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.888114 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-utilities" (OuterVolumeSpecName: "utilities") pod "0cf1d6a8-6356-4592-ae19-884f4b15bed6" (UID: "0cf1d6a8-6356-4592-ae19-884f4b15bed6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.892546 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cf1d6a8-6356-4592-ae19-884f4b15bed6-kube-api-access-bqw6p" (OuterVolumeSpecName: "kube-api-access-bqw6p") pod "0cf1d6a8-6356-4592-ae19-884f4b15bed6" (UID: "0cf1d6a8-6356-4592-ae19-884f4b15bed6"). InnerVolumeSpecName "kube-api-access-bqw6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.913492 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0cf1d6a8-6356-4592-ae19-884f4b15bed6" (UID: "0cf1d6a8-6356-4592-ae19-884f4b15bed6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.967184 4946 generic.go:334] "Generic (PLEG): container finished" podID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerID="b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5" exitCode=0 Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.967249 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5nr8" event={"ID":"0cf1d6a8-6356-4592-ae19-884f4b15bed6","Type":"ContainerDied","Data":"b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5"} Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.967293 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5nr8" event={"ID":"0cf1d6a8-6356-4592-ae19-884f4b15bed6","Type":"ContainerDied","Data":"53b25b5293ad6cb98b2ed72825a7e1d6fa1d81e5d7bacf40363e088c83e94b47"} Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.967321 4946 scope.go:117] "RemoveContainer" containerID="b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5" Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.967650 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5nr8" Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.988625 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.988656 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf1d6a8-6356-4592-ae19-884f4b15bed6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:17 crc kubenswrapper[4946]: I1203 07:35:17.988668 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqw6p\" (UniqueName: \"kubernetes.io/projected/0cf1d6a8-6356-4592-ae19-884f4b15bed6-kube-api-access-bqw6p\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.000936 4946 scope.go:117] "RemoveContainer" containerID="afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef" Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.003334 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5nr8"] Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.009267 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5nr8"] Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.038065 4946 scope.go:117] "RemoveContainer" containerID="0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750" Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.058044 4946 scope.go:117] "RemoveContainer" containerID="b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5" Dec 03 07:35:18 crc kubenswrapper[4946]: E1203 07:35:18.060316 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5\": container with ID starting with b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5 not found: ID does not exist" containerID="b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5" Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.060359 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5"} err="failed to get container status \"b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5\": rpc error: code = NotFound desc = could not find container \"b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5\": container with ID starting with b772603e48039cc7b011ae1ad0f4fb03a0d0a4cda19ccae4702439097af0d1e5 not found: ID does not exist" Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.060388 4946 scope.go:117] "RemoveContainer" containerID="afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef" Dec 03 07:35:18 crc kubenswrapper[4946]: E1203 07:35:18.060791 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef\": container with ID starting with afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef not found: ID does not exist" containerID="afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef" Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.060827 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef"} err="failed to get container status \"afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef\": rpc error: code = NotFound desc = could not find container \"afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef\": container with ID starting with afcd8491f8c3bfec0edbd664305f5f269bebcedccd9c6a5c90f08f9c71d7d5ef not found: ID does not exist" Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.060846 4946 scope.go:117] "RemoveContainer" containerID="0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750" Dec 03 07:35:18 crc kubenswrapper[4946]: E1203 07:35:18.061107 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750\": container with ID starting with 0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750 not found: ID does not exist" containerID="0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750" Dec 03 07:35:18 crc kubenswrapper[4946]: I1203 07:35:18.061133 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750"} err="failed to get container status \"0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750\": rpc error: code = NotFound desc = could not find container \"0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750\": container with ID starting with 0028056a53314a1a988a70c0abd304e0bb04aa4b2b2c6f4692eae1913dc33750 not found: ID does not exist" Dec 03 07:35:19 crc kubenswrapper[4946]: I1203 07:35:19.607988 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" path="/var/lib/kubelet/pods/0cf1d6a8-6356-4592-ae19-884f4b15bed6/volumes" Dec 03 07:35:23 crc kubenswrapper[4946]: I1203 07:35:23.040723 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:35:23 crc kubenswrapper[4946]: I1203 07:35:23.041210 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.787414 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vqvpk"] Dec 03 07:35:25 crc kubenswrapper[4946]: E1203 07:35:25.788318 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerName="extract-utilities" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.788340 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerName="extract-utilities" Dec 03 07:35:25 crc kubenswrapper[4946]: E1203 07:35:25.788382 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerName="extract-content" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.788395 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerName="extract-content" Dec 03 07:35:25 crc kubenswrapper[4946]: E1203 07:35:25.788415 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerName="registry-server" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.788428 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerName="registry-server" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.788781 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf1d6a8-6356-4592-ae19-884f4b15bed6" containerName="registry-server" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.790553 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.800163 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-utilities\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.800260 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-catalog-content\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.800349 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6vwj\" (UniqueName: \"kubernetes.io/projected/2e4128ae-91ff-4040-be29-c711f31a8144-kube-api-access-t6vwj\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.807462 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vqvpk"] Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.901472 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6vwj\" (UniqueName: \"kubernetes.io/projected/2e4128ae-91ff-4040-be29-c711f31a8144-kube-api-access-t6vwj\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.901829 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-utilities\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.901936 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-catalog-content\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.902231 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-utilities\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.902274 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-catalog-content\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:25 crc kubenswrapper[4946]: I1203 07:35:25.935445 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6vwj\" (UniqueName: \"kubernetes.io/projected/2e4128ae-91ff-4040-be29-c711f31a8144-kube-api-access-t6vwj\") pod \"certified-operators-vqvpk\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:26 crc kubenswrapper[4946]: I1203 07:35:26.137682 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:26 crc kubenswrapper[4946]: I1203 07:35:26.642027 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vqvpk"] Dec 03 07:35:27 crc kubenswrapper[4946]: I1203 07:35:27.044388 4946 generic.go:334] "Generic (PLEG): container finished" podID="2e4128ae-91ff-4040-be29-c711f31a8144" containerID="2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd" exitCode=0 Dec 03 07:35:27 crc kubenswrapper[4946]: I1203 07:35:27.044490 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqvpk" event={"ID":"2e4128ae-91ff-4040-be29-c711f31a8144","Type":"ContainerDied","Data":"2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd"} Dec 03 07:35:27 crc kubenswrapper[4946]: I1203 07:35:27.044908 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqvpk" event={"ID":"2e4128ae-91ff-4040-be29-c711f31a8144","Type":"ContainerStarted","Data":"7bc19f414e9d0c455b0e095da7c53f129656d67b34565d459ec90f5c39abeaa8"} Dec 03 07:35:28 crc kubenswrapper[4946]: I1203 07:35:28.053634 4946 generic.go:334] "Generic (PLEG): container finished" podID="2e4128ae-91ff-4040-be29-c711f31a8144" containerID="7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a" exitCode=0 Dec 03 07:35:28 crc kubenswrapper[4946]: I1203 07:35:28.053719 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqvpk" event={"ID":"2e4128ae-91ff-4040-be29-c711f31a8144","Type":"ContainerDied","Data":"7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a"} Dec 03 07:35:29 crc kubenswrapper[4946]: I1203 07:35:29.062916 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqvpk" event={"ID":"2e4128ae-91ff-4040-be29-c711f31a8144","Type":"ContainerStarted","Data":"a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da"} Dec 03 07:35:29 crc kubenswrapper[4946]: I1203 07:35:29.088141 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vqvpk" podStartSLOduration=2.6409649589999997 podStartE2EDuration="4.088120604s" podCreationTimestamp="2025-12-03 07:35:25 +0000 UTC" firstStartedPulling="2025-12-03 07:35:27.046061982 +0000 UTC m=+2719.842752091" lastFinishedPulling="2025-12-03 07:35:28.493217627 +0000 UTC m=+2721.289907736" observedRunningTime="2025-12-03 07:35:29.081546929 +0000 UTC m=+2721.878237038" watchObservedRunningTime="2025-12-03 07:35:29.088120604 +0000 UTC m=+2721.884810713" Dec 03 07:35:36 crc kubenswrapper[4946]: I1203 07:35:36.138040 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:36 crc kubenswrapper[4946]: I1203 07:35:36.138616 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:36 crc kubenswrapper[4946]: I1203 07:35:36.188184 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:37 crc kubenswrapper[4946]: I1203 07:35:37.162245 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:37 crc kubenswrapper[4946]: I1203 07:35:37.202423 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vqvpk"] Dec 03 07:35:39 crc kubenswrapper[4946]: I1203 07:35:39.142022 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vqvpk" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" containerName="registry-server" containerID="cri-o://a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da" gracePeriod=2 Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.093314 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.155187 4946 generic.go:334] "Generic (PLEG): container finished" podID="2e4128ae-91ff-4040-be29-c711f31a8144" containerID="a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da" exitCode=0 Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.155224 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqvpk" event={"ID":"2e4128ae-91ff-4040-be29-c711f31a8144","Type":"ContainerDied","Data":"a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da"} Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.155238 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqvpk" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.155249 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqvpk" event={"ID":"2e4128ae-91ff-4040-be29-c711f31a8144","Type":"ContainerDied","Data":"7bc19f414e9d0c455b0e095da7c53f129656d67b34565d459ec90f5c39abeaa8"} Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.155265 4946 scope.go:117] "RemoveContainer" containerID="a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.174707 4946 scope.go:117] "RemoveContainer" containerID="7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.190912 4946 scope.go:117] "RemoveContainer" containerID="2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.204956 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-catalog-content\") pod \"2e4128ae-91ff-4040-be29-c711f31a8144\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.205068 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6vwj\" (UniqueName: \"kubernetes.io/projected/2e4128ae-91ff-4040-be29-c711f31a8144-kube-api-access-t6vwj\") pod \"2e4128ae-91ff-4040-be29-c711f31a8144\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.205107 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-utilities\") pod \"2e4128ae-91ff-4040-be29-c711f31a8144\" (UID: \"2e4128ae-91ff-4040-be29-c711f31a8144\") " Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.208836 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-utilities" (OuterVolumeSpecName: "utilities") pod "2e4128ae-91ff-4040-be29-c711f31a8144" (UID: "2e4128ae-91ff-4040-be29-c711f31a8144"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.210712 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e4128ae-91ff-4040-be29-c711f31a8144-kube-api-access-t6vwj" (OuterVolumeSpecName: "kube-api-access-t6vwj") pod "2e4128ae-91ff-4040-be29-c711f31a8144" (UID: "2e4128ae-91ff-4040-be29-c711f31a8144"). InnerVolumeSpecName "kube-api-access-t6vwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.216218 4946 scope.go:117] "RemoveContainer" containerID="a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da" Dec 03 07:35:40 crc kubenswrapper[4946]: E1203 07:35:40.216622 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da\": container with ID starting with a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da not found: ID does not exist" containerID="a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.216672 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da"} err="failed to get container status \"a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da\": rpc error: code = NotFound desc = could not find container \"a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da\": container with ID starting with a62d1eb5f8924653ebf71dc85477a7cf1a26c51b1277e4a14b73054ec22ff6da not found: ID does not exist" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.216701 4946 scope.go:117] "RemoveContainer" containerID="7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a" Dec 03 07:35:40 crc kubenswrapper[4946]: E1203 07:35:40.217311 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a\": container with ID starting with 7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a not found: ID does not exist" containerID="7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.217353 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a"} err="failed to get container status \"7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a\": rpc error: code = NotFound desc = could not find container \"7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a\": container with ID starting with 7d2556df285f2234d603a8f8f6eeef8429802c92d91da84ef86950224673db4a not found: ID does not exist" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.217372 4946 scope.go:117] "RemoveContainer" containerID="2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd" Dec 03 07:35:40 crc kubenswrapper[4946]: E1203 07:35:40.217615 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd\": container with ID starting with 2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd not found: ID does not exist" containerID="2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.217646 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd"} err="failed to get container status \"2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd\": rpc error: code = NotFound desc = could not find container \"2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd\": container with ID starting with 2bfaf37a879506d64e365133749374e292a6488530e0e68ec24947d8c70872bd not found: ID does not exist" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.255170 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e4128ae-91ff-4040-be29-c711f31a8144" (UID: "2e4128ae-91ff-4040-be29-c711f31a8144"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.306394 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6vwj\" (UniqueName: \"kubernetes.io/projected/2e4128ae-91ff-4040-be29-c711f31a8144-kube-api-access-t6vwj\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.306436 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.306448 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e4128ae-91ff-4040-be29-c711f31a8144-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.490889 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vqvpk"] Dec 03 07:35:40 crc kubenswrapper[4946]: I1203 07:35:40.498183 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vqvpk"] Dec 03 07:35:41 crc kubenswrapper[4946]: I1203 07:35:41.605362 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" path="/var/lib/kubelet/pods/2e4128ae-91ff-4040-be29-c711f31a8144/volumes" Dec 03 07:35:53 crc kubenswrapper[4946]: I1203 07:35:53.039132 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:35:53 crc kubenswrapper[4946]: I1203 07:35:53.039787 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.443917 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-phngw"] Dec 03 07:36:13 crc kubenswrapper[4946]: E1203 07:36:13.445051 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" containerName="registry-server" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.445074 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" containerName="registry-server" Dec 03 07:36:13 crc kubenswrapper[4946]: E1203 07:36:13.445104 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" containerName="extract-utilities" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.445117 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" containerName="extract-utilities" Dec 03 07:36:13 crc kubenswrapper[4946]: E1203 07:36:13.445134 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" containerName="extract-content" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.445147 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" containerName="extract-content" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.445423 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e4128ae-91ff-4040-be29-c711f31a8144" containerName="registry-server" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.450445 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.484627 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-phngw"] Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.527859 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk7rc\" (UniqueName: \"kubernetes.io/projected/95058cda-3fe8-4ac7-9237-a3898f341e74-kube-api-access-gk7rc\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.527996 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-utilities\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.528041 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-catalog-content\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.629364 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk7rc\" (UniqueName: \"kubernetes.io/projected/95058cda-3fe8-4ac7-9237-a3898f341e74-kube-api-access-gk7rc\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.630652 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-utilities\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.630719 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-catalog-content\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.631573 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-utilities\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.632197 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-catalog-content\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.657418 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk7rc\" (UniqueName: \"kubernetes.io/projected/95058cda-3fe8-4ac7-9237-a3898f341e74-kube-api-access-gk7rc\") pod \"redhat-operators-phngw\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:13 crc kubenswrapper[4946]: I1203 07:36:13.792787 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:14 crc kubenswrapper[4946]: I1203 07:36:14.242031 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-phngw"] Dec 03 07:36:14 crc kubenswrapper[4946]: I1203 07:36:14.506040 4946 generic.go:334] "Generic (PLEG): container finished" podID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerID="3a02978934efb5994200bda8aceed2fa5df4b104c90521ca41c5d2fb3830afd0" exitCode=0 Dec 03 07:36:14 crc kubenswrapper[4946]: I1203 07:36:14.506104 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phngw" event={"ID":"95058cda-3fe8-4ac7-9237-a3898f341e74","Type":"ContainerDied","Data":"3a02978934efb5994200bda8aceed2fa5df4b104c90521ca41c5d2fb3830afd0"} Dec 03 07:36:14 crc kubenswrapper[4946]: I1203 07:36:14.506170 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phngw" event={"ID":"95058cda-3fe8-4ac7-9237-a3898f341e74","Type":"ContainerStarted","Data":"375a98e237352175d7c022c9596cd88c5279a09f80cdca88aa4d607d3d2e2a8a"} Dec 03 07:36:15 crc kubenswrapper[4946]: I1203 07:36:15.517334 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phngw" event={"ID":"95058cda-3fe8-4ac7-9237-a3898f341e74","Type":"ContainerStarted","Data":"e280a3e4325ca865bbe34591d8b7338c9a072c7f93218e82a2b19431c25d3ae7"} Dec 03 07:36:16 crc kubenswrapper[4946]: I1203 07:36:16.533673 4946 generic.go:334] "Generic (PLEG): container finished" podID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerID="e280a3e4325ca865bbe34591d8b7338c9a072c7f93218e82a2b19431c25d3ae7" exitCode=0 Dec 03 07:36:16 crc kubenswrapper[4946]: I1203 07:36:16.533795 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phngw" event={"ID":"95058cda-3fe8-4ac7-9237-a3898f341e74","Type":"ContainerDied","Data":"e280a3e4325ca865bbe34591d8b7338c9a072c7f93218e82a2b19431c25d3ae7"} Dec 03 07:36:17 crc kubenswrapper[4946]: I1203 07:36:17.542677 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phngw" event={"ID":"95058cda-3fe8-4ac7-9237-a3898f341e74","Type":"ContainerStarted","Data":"5ec61160c5425c7f4d89c76ca6a729865b8302d97dd1e6b052ec6151175b1e91"} Dec 03 07:36:17 crc kubenswrapper[4946]: I1203 07:36:17.576814 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-phngw" podStartSLOduration=2.170250621 podStartE2EDuration="4.576796563s" podCreationTimestamp="2025-12-03 07:36:13 +0000 UTC" firstStartedPulling="2025-12-03 07:36:14.508278303 +0000 UTC m=+2767.304968412" lastFinishedPulling="2025-12-03 07:36:16.914824205 +0000 UTC m=+2769.711514354" observedRunningTime="2025-12-03 07:36:17.568692337 +0000 UTC m=+2770.365382486" watchObservedRunningTime="2025-12-03 07:36:17.576796563 +0000 UTC m=+2770.373486672" Dec 03 07:36:23 crc kubenswrapper[4946]: I1203 07:36:23.039004 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:36:23 crc kubenswrapper[4946]: I1203 07:36:23.039645 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:36:23 crc kubenswrapper[4946]: I1203 07:36:23.039694 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:36:23 crc kubenswrapper[4946]: I1203 07:36:23.040318 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d78922732bd24992d7ce5ed423db2dc2eddfc0231ce52f4145d041f2ec118b6"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:36:23 crc kubenswrapper[4946]: I1203 07:36:23.040382 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://8d78922732bd24992d7ce5ed423db2dc2eddfc0231ce52f4145d041f2ec118b6" gracePeriod=600 Dec 03 07:36:23 crc kubenswrapper[4946]: I1203 07:36:23.793978 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:23 crc kubenswrapper[4946]: I1203 07:36:23.794941 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:23 crc kubenswrapper[4946]: I1203 07:36:23.855511 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:24 crc kubenswrapper[4946]: I1203 07:36:24.599267 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="8d78922732bd24992d7ce5ed423db2dc2eddfc0231ce52f4145d041f2ec118b6" exitCode=0 Dec 03 07:36:24 crc kubenswrapper[4946]: I1203 07:36:24.599342 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"8d78922732bd24992d7ce5ed423db2dc2eddfc0231ce52f4145d041f2ec118b6"} Dec 03 07:36:24 crc kubenswrapper[4946]: I1203 07:36:24.599704 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9"} Dec 03 07:36:24 crc kubenswrapper[4946]: I1203 07:36:24.599726 4946 scope.go:117] "RemoveContainer" containerID="b881215d257d6648e5e626a2efb480cb7e422752897717f4d082a991cbec80ad" Dec 03 07:36:24 crc kubenswrapper[4946]: I1203 07:36:24.644687 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:24 crc kubenswrapper[4946]: I1203 07:36:24.695828 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-phngw"] Dec 03 07:36:26 crc kubenswrapper[4946]: I1203 07:36:26.621152 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-phngw" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerName="registry-server" containerID="cri-o://5ec61160c5425c7f4d89c76ca6a729865b8302d97dd1e6b052ec6151175b1e91" gracePeriod=2 Dec 03 07:36:28 crc kubenswrapper[4946]: I1203 07:36:28.654908 4946 generic.go:334] "Generic (PLEG): container finished" podID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerID="5ec61160c5425c7f4d89c76ca6a729865b8302d97dd1e6b052ec6151175b1e91" exitCode=0 Dec 03 07:36:28 crc kubenswrapper[4946]: I1203 07:36:28.654978 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phngw" event={"ID":"95058cda-3fe8-4ac7-9237-a3898f341e74","Type":"ContainerDied","Data":"5ec61160c5425c7f4d89c76ca6a729865b8302d97dd1e6b052ec6151175b1e91"} Dec 03 07:36:28 crc kubenswrapper[4946]: I1203 07:36:28.928971 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:28 crc kubenswrapper[4946]: I1203 07:36:28.965198 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-utilities\") pod \"95058cda-3fe8-4ac7-9237-a3898f341e74\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " Dec 03 07:36:28 crc kubenswrapper[4946]: I1203 07:36:28.965318 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-catalog-content\") pod \"95058cda-3fe8-4ac7-9237-a3898f341e74\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " Dec 03 07:36:28 crc kubenswrapper[4946]: I1203 07:36:28.965375 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk7rc\" (UniqueName: \"kubernetes.io/projected/95058cda-3fe8-4ac7-9237-a3898f341e74-kube-api-access-gk7rc\") pod \"95058cda-3fe8-4ac7-9237-a3898f341e74\" (UID: \"95058cda-3fe8-4ac7-9237-a3898f341e74\") " Dec 03 07:36:28 crc kubenswrapper[4946]: I1203 07:36:28.966779 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-utilities" (OuterVolumeSpecName: "utilities") pod "95058cda-3fe8-4ac7-9237-a3898f341e74" (UID: "95058cda-3fe8-4ac7-9237-a3898f341e74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:36:28 crc kubenswrapper[4946]: I1203 07:36:28.973003 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95058cda-3fe8-4ac7-9237-a3898f341e74-kube-api-access-gk7rc" (OuterVolumeSpecName: "kube-api-access-gk7rc") pod "95058cda-3fe8-4ac7-9237-a3898f341e74" (UID: "95058cda-3fe8-4ac7-9237-a3898f341e74"). InnerVolumeSpecName "kube-api-access-gk7rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.067795 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk7rc\" (UniqueName: \"kubernetes.io/projected/95058cda-3fe8-4ac7-9237-a3898f341e74-kube-api-access-gk7rc\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.067898 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.079656 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95058cda-3fe8-4ac7-9237-a3898f341e74" (UID: "95058cda-3fe8-4ac7-9237-a3898f341e74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.168817 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95058cda-3fe8-4ac7-9237-a3898f341e74-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.668851 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phngw" event={"ID":"95058cda-3fe8-4ac7-9237-a3898f341e74","Type":"ContainerDied","Data":"375a98e237352175d7c022c9596cd88c5279a09f80cdca88aa4d607d3d2e2a8a"} Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.668909 4946 scope.go:117] "RemoveContainer" containerID="5ec61160c5425c7f4d89c76ca6a729865b8302d97dd1e6b052ec6151175b1e91" Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.668943 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phngw" Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.692372 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-phngw"] Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.699173 4946 scope.go:117] "RemoveContainer" containerID="e280a3e4325ca865bbe34591d8b7338c9a072c7f93218e82a2b19431c25d3ae7" Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.701697 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-phngw"] Dec 03 07:36:29 crc kubenswrapper[4946]: I1203 07:36:29.724213 4946 scope.go:117] "RemoveContainer" containerID="3a02978934efb5994200bda8aceed2fa5df4b104c90521ca41c5d2fb3830afd0" Dec 03 07:36:31 crc kubenswrapper[4946]: I1203 07:36:31.603726 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" path="/var/lib/kubelet/pods/95058cda-3fe8-4ac7-9237-a3898f341e74/volumes" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.107430 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gtn8p"] Dec 03 07:38:34 crc kubenswrapper[4946]: E1203 07:38:34.115032 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerName="extract-utilities" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.115076 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerName="extract-utilities" Dec 03 07:38:34 crc kubenswrapper[4946]: E1203 07:38:34.115090 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerName="registry-server" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.115102 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerName="registry-server" Dec 03 07:38:34 crc kubenswrapper[4946]: E1203 07:38:34.115125 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerName="extract-content" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.115138 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerName="extract-content" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.115365 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="95058cda-3fe8-4ac7-9237-a3898f341e74" containerName="registry-server" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.116872 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.159490 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtn8p"] Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.258805 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvs4n\" (UniqueName: \"kubernetes.io/projected/6ca0020d-490e-4efe-a145-a4bee0006e53-kube-api-access-gvs4n\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.258894 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ca0020d-490e-4efe-a145-a4bee0006e53-utilities\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.258985 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ca0020d-490e-4efe-a145-a4bee0006e53-catalog-content\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.359944 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ca0020d-490e-4efe-a145-a4bee0006e53-catalog-content\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.360025 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvs4n\" (UniqueName: \"kubernetes.io/projected/6ca0020d-490e-4efe-a145-a4bee0006e53-kube-api-access-gvs4n\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.360070 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ca0020d-490e-4efe-a145-a4bee0006e53-utilities\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.360499 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ca0020d-490e-4efe-a145-a4bee0006e53-catalog-content\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.360509 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ca0020d-490e-4efe-a145-a4bee0006e53-utilities\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.383853 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvs4n\" (UniqueName: \"kubernetes.io/projected/6ca0020d-490e-4efe-a145-a4bee0006e53-kube-api-access-gvs4n\") pod \"community-operators-gtn8p\" (UID: \"6ca0020d-490e-4efe-a145-a4bee0006e53\") " pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.460043 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:34 crc kubenswrapper[4946]: I1203 07:38:34.929655 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtn8p"] Dec 03 07:38:34 crc kubenswrapper[4946]: W1203 07:38:34.938576 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ca0020d_490e_4efe_a145_a4bee0006e53.slice/crio-60b8fcf59e2b20cb94c39e552d559dd913663d7b2aa28be77fd87038d4ed1e40 WatchSource:0}: Error finding container 60b8fcf59e2b20cb94c39e552d559dd913663d7b2aa28be77fd87038d4ed1e40: Status 404 returned error can't find the container with id 60b8fcf59e2b20cb94c39e552d559dd913663d7b2aa28be77fd87038d4ed1e40 Dec 03 07:38:35 crc kubenswrapper[4946]: I1203 07:38:35.821120 4946 generic.go:334] "Generic (PLEG): container finished" podID="6ca0020d-490e-4efe-a145-a4bee0006e53" containerID="65a80ec79a8135441cc45c8fea75765b48ef4146f567af9bedb5d079f8560de0" exitCode=0 Dec 03 07:38:35 crc kubenswrapper[4946]: I1203 07:38:35.821161 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtn8p" event={"ID":"6ca0020d-490e-4efe-a145-a4bee0006e53","Type":"ContainerDied","Data":"65a80ec79a8135441cc45c8fea75765b48ef4146f567af9bedb5d079f8560de0"} Dec 03 07:38:35 crc kubenswrapper[4946]: I1203 07:38:35.821187 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtn8p" event={"ID":"6ca0020d-490e-4efe-a145-a4bee0006e53","Type":"ContainerStarted","Data":"60b8fcf59e2b20cb94c39e552d559dd913663d7b2aa28be77fd87038d4ed1e40"} Dec 03 07:38:39 crc kubenswrapper[4946]: I1203 07:38:39.857991 4946 generic.go:334] "Generic (PLEG): container finished" podID="6ca0020d-490e-4efe-a145-a4bee0006e53" containerID="4d1fb6f34ce2dcba2cffcd42a8740f440f570922093e60b3fd9722144ee55e33" exitCode=0 Dec 03 07:38:39 crc kubenswrapper[4946]: I1203 07:38:39.858071 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtn8p" event={"ID":"6ca0020d-490e-4efe-a145-a4bee0006e53","Type":"ContainerDied","Data":"4d1fb6f34ce2dcba2cffcd42a8740f440f570922093e60b3fd9722144ee55e33"} Dec 03 07:38:40 crc kubenswrapper[4946]: I1203 07:38:40.872966 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtn8p" event={"ID":"6ca0020d-490e-4efe-a145-a4bee0006e53","Type":"ContainerStarted","Data":"12b90b5b722183a7110ab51d9360a90523110f62a0109f5bd8c7b157e00380fc"} Dec 03 07:38:40 crc kubenswrapper[4946]: I1203 07:38:40.897485 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gtn8p" podStartSLOduration=2.44032548 podStartE2EDuration="6.897465933s" podCreationTimestamp="2025-12-03 07:38:34 +0000 UTC" firstStartedPulling="2025-12-03 07:38:35.824232556 +0000 UTC m=+2908.620922665" lastFinishedPulling="2025-12-03 07:38:40.281372999 +0000 UTC m=+2913.078063118" observedRunningTime="2025-12-03 07:38:40.89284355 +0000 UTC m=+2913.689533699" watchObservedRunningTime="2025-12-03 07:38:40.897465933 +0000 UTC m=+2913.694156062" Dec 03 07:38:44 crc kubenswrapper[4946]: I1203 07:38:44.461013 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:44 crc kubenswrapper[4946]: I1203 07:38:44.462092 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:44 crc kubenswrapper[4946]: I1203 07:38:44.533899 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:53 crc kubenswrapper[4946]: I1203 07:38:53.039395 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:38:53 crc kubenswrapper[4946]: I1203 07:38:53.040208 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:38:54 crc kubenswrapper[4946]: I1203 07:38:54.530603 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gtn8p" Dec 03 07:38:54 crc kubenswrapper[4946]: I1203 07:38:54.633499 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtn8p"] Dec 03 07:38:54 crc kubenswrapper[4946]: I1203 07:38:54.682480 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tww9n"] Dec 03 07:38:54 crc kubenswrapper[4946]: I1203 07:38:54.683046 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tww9n" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerName="registry-server" containerID="cri-o://123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752" gracePeriod=2 Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.000132 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tww9n" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.009616 4946 generic.go:334] "Generic (PLEG): container finished" podID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerID="123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752" exitCode=0 Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.009653 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tww9n" event={"ID":"95456701-a430-44aa-9f6a-cf6daeb3e0e0","Type":"ContainerDied","Data":"123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752"} Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.009681 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tww9n" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.009692 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tww9n" event={"ID":"95456701-a430-44aa-9f6a-cf6daeb3e0e0","Type":"ContainerDied","Data":"81303a5926d8ccf7dc25a61404a6c5e45e4a7505ed42957df55557a6f0590bfe"} Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.009712 4946 scope.go:117] "RemoveContainer" containerID="123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.043632 4946 scope.go:117] "RemoveContainer" containerID="2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.077647 4946 scope.go:117] "RemoveContainer" containerID="5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.104552 4946 scope.go:117] "RemoveContainer" containerID="123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752" Dec 03 07:38:57 crc kubenswrapper[4946]: E1203 07:38:57.105013 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752\": container with ID starting with 123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752 not found: ID does not exist" containerID="123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.105051 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752"} err="failed to get container status \"123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752\": rpc error: code = NotFound desc = could not find container \"123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752\": container with ID starting with 123750508df3523356251c7b4431310fb9e950c6305b4b3af3c3c425c6723752 not found: ID does not exist" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.105072 4946 scope.go:117] "RemoveContainer" containerID="2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8" Dec 03 07:38:57 crc kubenswrapper[4946]: E1203 07:38:57.106101 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8\": container with ID starting with 2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8 not found: ID does not exist" containerID="2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.106126 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8"} err="failed to get container status \"2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8\": rpc error: code = NotFound desc = could not find container \"2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8\": container with ID starting with 2910ccbf56feba350d85f559643a6e0edead4fc86c4c8e18a586f0563c1a98a8 not found: ID does not exist" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.106138 4946 scope.go:117] "RemoveContainer" containerID="5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f" Dec 03 07:38:57 crc kubenswrapper[4946]: E1203 07:38:57.106648 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f\": container with ID starting with 5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f not found: ID does not exist" containerID="5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.106670 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f"} err="failed to get container status \"5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f\": rpc error: code = NotFound desc = could not find container \"5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f\": container with ID starting with 5079f5adb3b8c43749574b78da4159789b081e477fd8ed529c5191c0daf9798f not found: ID does not exist" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.118289 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb5xz\" (UniqueName: \"kubernetes.io/projected/95456701-a430-44aa-9f6a-cf6daeb3e0e0-kube-api-access-xb5xz\") pod \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.118387 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-catalog-content\") pod \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.118447 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-utilities\") pod \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\" (UID: \"95456701-a430-44aa-9f6a-cf6daeb3e0e0\") " Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.119520 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-utilities" (OuterVolumeSpecName: "utilities") pod "95456701-a430-44aa-9f6a-cf6daeb3e0e0" (UID: "95456701-a430-44aa-9f6a-cf6daeb3e0e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.124989 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95456701-a430-44aa-9f6a-cf6daeb3e0e0-kube-api-access-xb5xz" (OuterVolumeSpecName: "kube-api-access-xb5xz") pod "95456701-a430-44aa-9f6a-cf6daeb3e0e0" (UID: "95456701-a430-44aa-9f6a-cf6daeb3e0e0"). InnerVolumeSpecName "kube-api-access-xb5xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.184510 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95456701-a430-44aa-9f6a-cf6daeb3e0e0" (UID: "95456701-a430-44aa-9f6a-cf6daeb3e0e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.221325 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb5xz\" (UniqueName: \"kubernetes.io/projected/95456701-a430-44aa-9f6a-cf6daeb3e0e0-kube-api-access-xb5xz\") on node \"crc\" DevicePath \"\"" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.221375 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.221389 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95456701-a430-44aa-9f6a-cf6daeb3e0e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.339143 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tww9n"] Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.343581 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tww9n"] Dec 03 07:38:57 crc kubenswrapper[4946]: I1203 07:38:57.600951 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" path="/var/lib/kubelet/pods/95456701-a430-44aa-9f6a-cf6daeb3e0e0/volumes" Dec 03 07:39:23 crc kubenswrapper[4946]: I1203 07:39:23.039838 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:39:23 crc kubenswrapper[4946]: I1203 07:39:23.040404 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.040020 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.040861 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.040931 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.041795 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.041906 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" gracePeriod=600 Dec 03 07:39:53 crc kubenswrapper[4946]: E1203 07:39:53.167644 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.489667 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" exitCode=0 Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.489721 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9"} Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.489890 4946 scope.go:117] "RemoveContainer" containerID="8d78922732bd24992d7ce5ed423db2dc2eddfc0231ce52f4145d041f2ec118b6" Dec 03 07:39:53 crc kubenswrapper[4946]: I1203 07:39:53.490876 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:39:53 crc kubenswrapper[4946]: E1203 07:39:53.491516 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:40:04 crc kubenswrapper[4946]: I1203 07:40:04.593350 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:40:04 crc kubenswrapper[4946]: E1203 07:40:04.594229 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:40:15 crc kubenswrapper[4946]: I1203 07:40:15.595550 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:40:15 crc kubenswrapper[4946]: E1203 07:40:15.597177 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:40:30 crc kubenswrapper[4946]: I1203 07:40:30.592987 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:40:30 crc kubenswrapper[4946]: E1203 07:40:30.594101 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:40:41 crc kubenswrapper[4946]: I1203 07:40:41.592862 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:40:41 crc kubenswrapper[4946]: E1203 07:40:41.593686 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:40:56 crc kubenswrapper[4946]: I1203 07:40:56.593639 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:40:56 crc kubenswrapper[4946]: E1203 07:40:56.594822 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:41:08 crc kubenswrapper[4946]: I1203 07:41:08.593204 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:41:08 crc kubenswrapper[4946]: E1203 07:41:08.594248 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:41:19 crc kubenswrapper[4946]: I1203 07:41:19.592602 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:41:19 crc kubenswrapper[4946]: E1203 07:41:19.593378 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:41:34 crc kubenswrapper[4946]: I1203 07:41:34.592369 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:41:34 crc kubenswrapper[4946]: E1203 07:41:34.593119 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:41:48 crc kubenswrapper[4946]: I1203 07:41:48.592729 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:41:48 crc kubenswrapper[4946]: E1203 07:41:48.593803 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:42:02 crc kubenswrapper[4946]: I1203 07:42:02.593712 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:42:02 crc kubenswrapper[4946]: E1203 07:42:02.595119 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:42:14 crc kubenswrapper[4946]: I1203 07:42:14.593064 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:42:14 crc kubenswrapper[4946]: E1203 07:42:14.593852 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:42:28 crc kubenswrapper[4946]: I1203 07:42:28.592719 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:42:28 crc kubenswrapper[4946]: E1203 07:42:28.594007 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:42:43 crc kubenswrapper[4946]: I1203 07:42:43.593519 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:42:43 crc kubenswrapper[4946]: E1203 07:42:43.594436 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:42:54 crc kubenswrapper[4946]: I1203 07:42:54.592694 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:42:54 crc kubenswrapper[4946]: E1203 07:42:54.593526 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:43:06 crc kubenswrapper[4946]: I1203 07:43:06.593904 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:43:06 crc kubenswrapper[4946]: E1203 07:43:06.594926 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:43:17 crc kubenswrapper[4946]: I1203 07:43:17.598021 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:43:17 crc kubenswrapper[4946]: E1203 07:43:17.598985 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:43:30 crc kubenswrapper[4946]: I1203 07:43:30.592987 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:43:30 crc kubenswrapper[4946]: E1203 07:43:30.594191 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:43:44 crc kubenswrapper[4946]: I1203 07:43:44.592557 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:43:44 crc kubenswrapper[4946]: E1203 07:43:44.593401 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:43:59 crc kubenswrapper[4946]: I1203 07:43:59.592376 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:43:59 crc kubenswrapper[4946]: E1203 07:43:59.593320 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:44:11 crc kubenswrapper[4946]: I1203 07:44:11.593668 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:44:11 crc kubenswrapper[4946]: E1203 07:44:11.594421 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:44:23 crc kubenswrapper[4946]: I1203 07:44:23.594219 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:44:23 crc kubenswrapper[4946]: E1203 07:44:23.595447 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:44:36 crc kubenswrapper[4946]: I1203 07:44:36.593031 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:44:36 crc kubenswrapper[4946]: E1203 07:44:36.594105 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:44:50 crc kubenswrapper[4946]: I1203 07:44:50.594017 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:44:50 crc kubenswrapper[4946]: E1203 07:44:50.595268 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.169939 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd"] Dec 03 07:45:00 crc kubenswrapper[4946]: E1203 07:45:00.171008 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerName="registry-server" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.171030 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerName="registry-server" Dec 03 07:45:00 crc kubenswrapper[4946]: E1203 07:45:00.171071 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerName="extract-content" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.171080 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerName="extract-content" Dec 03 07:45:00 crc kubenswrapper[4946]: E1203 07:45:00.171103 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerName="extract-utilities" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.171112 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerName="extract-utilities" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.171300 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="95456701-a430-44aa-9f6a-cf6daeb3e0e0" containerName="registry-server" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.171975 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.176208 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.176940 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.184481 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd"] Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.303781 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-secret-volume\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.303880 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdmbk\" (UniqueName: \"kubernetes.io/projected/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-kube-api-access-tdmbk\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.303972 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-config-volume\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.406020 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-secret-volume\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.406095 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdmbk\" (UniqueName: \"kubernetes.io/projected/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-kube-api-access-tdmbk\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.406168 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-config-volume\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.408769 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-config-volume\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.416205 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-secret-volume\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.423594 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdmbk\" (UniqueName: \"kubernetes.io/projected/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-kube-api-access-tdmbk\") pod \"collect-profiles-29412465-j59sd\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.499721 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:00 crc kubenswrapper[4946]: I1203 07:45:00.949098 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd"] Dec 03 07:45:01 crc kubenswrapper[4946]: I1203 07:45:01.279649 4946 generic.go:334] "Generic (PLEG): container finished" podID="24c5cc9e-1a0d-4d92-9f86-61684d69f75d" containerID="f39203099fbc636700244e31c85f7af3078fad13445b07d580c81ca54ef64cbe" exitCode=0 Dec 03 07:45:01 crc kubenswrapper[4946]: I1203 07:45:01.279784 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" event={"ID":"24c5cc9e-1a0d-4d92-9f86-61684d69f75d","Type":"ContainerDied","Data":"f39203099fbc636700244e31c85f7af3078fad13445b07d580c81ca54ef64cbe"} Dec 03 07:45:01 crc kubenswrapper[4946]: I1203 07:45:01.279996 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" event={"ID":"24c5cc9e-1a0d-4d92-9f86-61684d69f75d","Type":"ContainerStarted","Data":"7e5110887c9ee7ac7dd08c6bdb7ae1ef29b6a0071761989cc4e8b58d9d75f5ec"} Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.594424 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.597436 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.749081 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdmbk\" (UniqueName: \"kubernetes.io/projected/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-kube-api-access-tdmbk\") pod \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.749339 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-config-volume\") pod \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.749395 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-secret-volume\") pod \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\" (UID: \"24c5cc9e-1a0d-4d92-9f86-61684d69f75d\") " Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.749955 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-config-volume" (OuterVolumeSpecName: "config-volume") pod "24c5cc9e-1a0d-4d92-9f86-61684d69f75d" (UID: "24c5cc9e-1a0d-4d92-9f86-61684d69f75d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.753531 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "24c5cc9e-1a0d-4d92-9f86-61684d69f75d" (UID: "24c5cc9e-1a0d-4d92-9f86-61684d69f75d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.759525 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-kube-api-access-tdmbk" (OuterVolumeSpecName: "kube-api-access-tdmbk") pod "24c5cc9e-1a0d-4d92-9f86-61684d69f75d" (UID: "24c5cc9e-1a0d-4d92-9f86-61684d69f75d"). InnerVolumeSpecName "kube-api-access-tdmbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.851073 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.851113 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdmbk\" (UniqueName: \"kubernetes.io/projected/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-kube-api-access-tdmbk\") on node \"crc\" DevicePath \"\"" Dec 03 07:45:02 crc kubenswrapper[4946]: I1203 07:45:02.851123 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24c5cc9e-1a0d-4d92-9f86-61684d69f75d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:45:03 crc kubenswrapper[4946]: I1203 07:45:03.305239 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" Dec 03 07:45:03 crc kubenswrapper[4946]: I1203 07:45:03.305733 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd" event={"ID":"24c5cc9e-1a0d-4d92-9f86-61684d69f75d","Type":"ContainerDied","Data":"7e5110887c9ee7ac7dd08c6bdb7ae1ef29b6a0071761989cc4e8b58d9d75f5ec"} Dec 03 07:45:03 crc kubenswrapper[4946]: I1203 07:45:03.305821 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e5110887c9ee7ac7dd08c6bdb7ae1ef29b6a0071761989cc4e8b58d9d75f5ec" Dec 03 07:45:03 crc kubenswrapper[4946]: I1203 07:45:03.308531 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"ae1d7773e3ac40586172ec68cf3a08b2f4a5181ad947298f33fcaa0b2cb770dc"} Dec 03 07:45:03 crc kubenswrapper[4946]: I1203 07:45:03.698037 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc"] Dec 03 07:45:03 crc kubenswrapper[4946]: I1203 07:45:03.704278 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412420-zj9cc"] Dec 03 07:45:05 crc kubenswrapper[4946]: I1203 07:45:05.607357 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b1cf6c8-022c-4f2b-b91f-4472579546cb" path="/var/lib/kubelet/pods/2b1cf6c8-022c-4f2b-b91f-4472579546cb/volumes" Dec 03 07:45:13 crc kubenswrapper[4946]: I1203 07:45:13.361312 4946 scope.go:117] "RemoveContainer" containerID="5098e930dbe5b23b7e8b9f2804444be8d5fbeba493e0d2ffa76e15a93d2130fe" Dec 03 07:46:40 crc kubenswrapper[4946]: I1203 07:46:40.795611 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8nvl7"] Dec 03 07:46:40 crc kubenswrapper[4946]: E1203 07:46:40.796364 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c5cc9e-1a0d-4d92-9f86-61684d69f75d" containerName="collect-profiles" Dec 03 07:46:40 crc kubenswrapper[4946]: I1203 07:46:40.796377 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c5cc9e-1a0d-4d92-9f86-61684d69f75d" containerName="collect-profiles" Dec 03 07:46:40 crc kubenswrapper[4946]: I1203 07:46:40.796533 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c5cc9e-1a0d-4d92-9f86-61684d69f75d" containerName="collect-profiles" Dec 03 07:46:40 crc kubenswrapper[4946]: I1203 07:46:40.797441 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:40 crc kubenswrapper[4946]: I1203 07:46:40.812068 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8nvl7"] Dec 03 07:46:40 crc kubenswrapper[4946]: I1203 07:46:40.908113 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-catalog-content\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:40 crc kubenswrapper[4946]: I1203 07:46:40.908156 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkght\" (UniqueName: \"kubernetes.io/projected/93b3f133-46dd-4fc2-949d-998c7d4687fb-kube-api-access-pkght\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:40 crc kubenswrapper[4946]: I1203 07:46:40.908193 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-utilities\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:41 crc kubenswrapper[4946]: I1203 07:46:41.009399 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-utilities\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:41 crc kubenswrapper[4946]: I1203 07:46:41.009978 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-utilities\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:41 crc kubenswrapper[4946]: I1203 07:46:41.010637 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-catalog-content\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:41 crc kubenswrapper[4946]: I1203 07:46:41.011232 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkght\" (UniqueName: \"kubernetes.io/projected/93b3f133-46dd-4fc2-949d-998c7d4687fb-kube-api-access-pkght\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:41 crc kubenswrapper[4946]: I1203 07:46:41.011376 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-catalog-content\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:41 crc kubenswrapper[4946]: I1203 07:46:41.034984 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkght\" (UniqueName: \"kubernetes.io/projected/93b3f133-46dd-4fc2-949d-998c7d4687fb-kube-api-access-pkght\") pod \"certified-operators-8nvl7\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:41 crc kubenswrapper[4946]: I1203 07:46:41.168448 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:41 crc kubenswrapper[4946]: I1203 07:46:41.696037 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8nvl7"] Dec 03 07:46:42 crc kubenswrapper[4946]: I1203 07:46:42.211494 4946 generic.go:334] "Generic (PLEG): container finished" podID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerID="39570ec2280d14a5cd4b6b67a50933d31f7afd247f7b13d62cef603d129b4b51" exitCode=0 Dec 03 07:46:42 crc kubenswrapper[4946]: I1203 07:46:42.211557 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nvl7" event={"ID":"93b3f133-46dd-4fc2-949d-998c7d4687fb","Type":"ContainerDied","Data":"39570ec2280d14a5cd4b6b67a50933d31f7afd247f7b13d62cef603d129b4b51"} Dec 03 07:46:42 crc kubenswrapper[4946]: I1203 07:46:42.211599 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nvl7" event={"ID":"93b3f133-46dd-4fc2-949d-998c7d4687fb","Type":"ContainerStarted","Data":"ea343f7871a6f9526a349931761c89655d38d65275e50041791a41692e4ba0b3"} Dec 03 07:46:42 crc kubenswrapper[4946]: I1203 07:46:42.214408 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:46:43 crc kubenswrapper[4946]: I1203 07:46:43.220421 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nvl7" event={"ID":"93b3f133-46dd-4fc2-949d-998c7d4687fb","Type":"ContainerStarted","Data":"4b5ce0dc265ac7d55c74bb1f668228d4605345e5ceb9fb1bf28ebe8286b108f0"} Dec 03 07:46:44 crc kubenswrapper[4946]: I1203 07:46:44.229002 4946 generic.go:334] "Generic (PLEG): container finished" podID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerID="4b5ce0dc265ac7d55c74bb1f668228d4605345e5ceb9fb1bf28ebe8286b108f0" exitCode=0 Dec 03 07:46:44 crc kubenswrapper[4946]: I1203 07:46:44.229063 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nvl7" event={"ID":"93b3f133-46dd-4fc2-949d-998c7d4687fb","Type":"ContainerDied","Data":"4b5ce0dc265ac7d55c74bb1f668228d4605345e5ceb9fb1bf28ebe8286b108f0"} Dec 03 07:46:45 crc kubenswrapper[4946]: I1203 07:46:45.244508 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nvl7" event={"ID":"93b3f133-46dd-4fc2-949d-998c7d4687fb","Type":"ContainerStarted","Data":"ae23ff13e81cfc7d31e64a12984de78ee7cae279d5bd7c6e45b7fead7f094519"} Dec 03 07:46:45 crc kubenswrapper[4946]: I1203 07:46:45.281514 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8nvl7" podStartSLOduration=2.907183476 podStartE2EDuration="5.281480678s" podCreationTimestamp="2025-12-03 07:46:40 +0000 UTC" firstStartedPulling="2025-12-03 07:46:42.214050176 +0000 UTC m=+3395.010740325" lastFinishedPulling="2025-12-03 07:46:44.588347378 +0000 UTC m=+3397.385037527" observedRunningTime="2025-12-03 07:46:45.27143811 +0000 UTC m=+3398.068128259" watchObservedRunningTime="2025-12-03 07:46:45.281480678 +0000 UTC m=+3398.078170827" Dec 03 07:46:51 crc kubenswrapper[4946]: I1203 07:46:51.169728 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:51 crc kubenswrapper[4946]: I1203 07:46:51.170355 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:51 crc kubenswrapper[4946]: I1203 07:46:51.246987 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:51 crc kubenswrapper[4946]: I1203 07:46:51.373059 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:51 crc kubenswrapper[4946]: I1203 07:46:51.491669 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8nvl7"] Dec 03 07:46:53 crc kubenswrapper[4946]: I1203 07:46:53.322092 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8nvl7" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerName="registry-server" containerID="cri-o://ae23ff13e81cfc7d31e64a12984de78ee7cae279d5bd7c6e45b7fead7f094519" gracePeriod=2 Dec 03 07:46:54 crc kubenswrapper[4946]: I1203 07:46:54.332420 4946 generic.go:334] "Generic (PLEG): container finished" podID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerID="ae23ff13e81cfc7d31e64a12984de78ee7cae279d5bd7c6e45b7fead7f094519" exitCode=0 Dec 03 07:46:54 crc kubenswrapper[4946]: I1203 07:46:54.332460 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nvl7" event={"ID":"93b3f133-46dd-4fc2-949d-998c7d4687fb","Type":"ContainerDied","Data":"ae23ff13e81cfc7d31e64a12984de78ee7cae279d5bd7c6e45b7fead7f094519"} Dec 03 07:46:54 crc kubenswrapper[4946]: I1203 07:46:54.915952 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.029960 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-catalog-content\") pod \"93b3f133-46dd-4fc2-949d-998c7d4687fb\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.030455 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkght\" (UniqueName: \"kubernetes.io/projected/93b3f133-46dd-4fc2-949d-998c7d4687fb-kube-api-access-pkght\") pod \"93b3f133-46dd-4fc2-949d-998c7d4687fb\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.030529 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-utilities\") pod \"93b3f133-46dd-4fc2-949d-998c7d4687fb\" (UID: \"93b3f133-46dd-4fc2-949d-998c7d4687fb\") " Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.031497 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-utilities" (OuterVolumeSpecName: "utilities") pod "93b3f133-46dd-4fc2-949d-998c7d4687fb" (UID: "93b3f133-46dd-4fc2-949d-998c7d4687fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.031651 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.043222 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93b3f133-46dd-4fc2-949d-998c7d4687fb-kube-api-access-pkght" (OuterVolumeSpecName: "kube-api-access-pkght") pod "93b3f133-46dd-4fc2-949d-998c7d4687fb" (UID: "93b3f133-46dd-4fc2-949d-998c7d4687fb"). InnerVolumeSpecName "kube-api-access-pkght". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.094922 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93b3f133-46dd-4fc2-949d-998c7d4687fb" (UID: "93b3f133-46dd-4fc2-949d-998c7d4687fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.132802 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkght\" (UniqueName: \"kubernetes.io/projected/93b3f133-46dd-4fc2-949d-998c7d4687fb-kube-api-access-pkght\") on node \"crc\" DevicePath \"\"" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.132837 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93b3f133-46dd-4fc2-949d-998c7d4687fb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.342415 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8nvl7" event={"ID":"93b3f133-46dd-4fc2-949d-998c7d4687fb","Type":"ContainerDied","Data":"ea343f7871a6f9526a349931761c89655d38d65275e50041791a41692e4ba0b3"} Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.342505 4946 scope.go:117] "RemoveContainer" containerID="ae23ff13e81cfc7d31e64a12984de78ee7cae279d5bd7c6e45b7fead7f094519" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.342519 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8nvl7" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.396452 4946 scope.go:117] "RemoveContainer" containerID="4b5ce0dc265ac7d55c74bb1f668228d4605345e5ceb9fb1bf28ebe8286b108f0" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.421728 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8nvl7"] Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.436275 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8nvl7"] Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.444270 4946 scope.go:117] "RemoveContainer" containerID="39570ec2280d14a5cd4b6b67a50933d31f7afd247f7b13d62cef603d129b4b51" Dec 03 07:46:55 crc kubenswrapper[4946]: I1203 07:46:55.602993 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" path="/var/lib/kubelet/pods/93b3f133-46dd-4fc2-949d-998c7d4687fb/volumes" Dec 03 07:47:23 crc kubenswrapper[4946]: I1203 07:47:23.039346 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:47:23 crc kubenswrapper[4946]: I1203 07:47:23.040128 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:47:53 crc kubenswrapper[4946]: I1203 07:47:53.039160 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:47:53 crc kubenswrapper[4946]: I1203 07:47:53.040095 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:48:23 crc kubenswrapper[4946]: I1203 07:48:23.039710 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:48:23 crc kubenswrapper[4946]: I1203 07:48:23.040571 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:48:23 crc kubenswrapper[4946]: I1203 07:48:23.040643 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:48:23 crc kubenswrapper[4946]: I1203 07:48:23.041583 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae1d7773e3ac40586172ec68cf3a08b2f4a5181ad947298f33fcaa0b2cb770dc"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:48:23 crc kubenswrapper[4946]: I1203 07:48:23.041705 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://ae1d7773e3ac40586172ec68cf3a08b2f4a5181ad947298f33fcaa0b2cb770dc" gracePeriod=600 Dec 03 07:48:23 crc kubenswrapper[4946]: I1203 07:48:23.212335 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="ae1d7773e3ac40586172ec68cf3a08b2f4a5181ad947298f33fcaa0b2cb770dc" exitCode=0 Dec 03 07:48:23 crc kubenswrapper[4946]: I1203 07:48:23.212451 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"ae1d7773e3ac40586172ec68cf3a08b2f4a5181ad947298f33fcaa0b2cb770dc"} Dec 03 07:48:23 crc kubenswrapper[4946]: I1203 07:48:23.212952 4946 scope.go:117] "RemoveContainer" containerID="95ab8c9c9b50fc59c3da945fe2bec4fdd55b719f79184ff27d50774afaed35e9" Dec 03 07:48:24 crc kubenswrapper[4946]: I1203 07:48:24.224725 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9"} Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.870375 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-25h7x"] Dec 03 07:49:10 crc kubenswrapper[4946]: E1203 07:49:10.871537 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerName="extract-utilities" Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.871559 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerName="extract-utilities" Dec 03 07:49:10 crc kubenswrapper[4946]: E1203 07:49:10.871591 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerName="extract-content" Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.871604 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerName="extract-content" Dec 03 07:49:10 crc kubenswrapper[4946]: E1203 07:49:10.871624 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerName="registry-server" Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.871637 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerName="registry-server" Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.871932 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="93b3f133-46dd-4fc2-949d-998c7d4687fb" containerName="registry-server" Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.873769 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.912296 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25h7x"] Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.955325 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-catalog-content\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.955398 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j894v\" (UniqueName: \"kubernetes.io/projected/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-kube-api-access-j894v\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:10 crc kubenswrapper[4946]: I1203 07:49:10.955626 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-utilities\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.057521 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-utilities\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.057677 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-catalog-content\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.057760 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j894v\" (UniqueName: \"kubernetes.io/projected/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-kube-api-access-j894v\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.058175 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-utilities\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.058265 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-catalog-content\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.092236 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j894v\" (UniqueName: \"kubernetes.io/projected/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-kube-api-access-j894v\") pod \"redhat-marketplace-25h7x\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.204249 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.460157 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sdl64"] Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.462568 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.469241 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25h7x"] Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.478269 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdl64"] Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.564337 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-utilities\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.564451 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-catalog-content\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.564501 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt6th\" (UniqueName: \"kubernetes.io/projected/98f8e837-97e0-48fd-a9ab-8383184a052c-kube-api-access-kt6th\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.665727 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-catalog-content\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.665831 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt6th\" (UniqueName: \"kubernetes.io/projected/98f8e837-97e0-48fd-a9ab-8383184a052c-kube-api-access-kt6th\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.665917 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-utilities\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.666187 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-catalog-content\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.666250 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-utilities\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.682732 4946 generic.go:334] "Generic (PLEG): container finished" podID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerID="2c260652bebf9782a63a6ac5921b70138d778e4b69d3f500bdb317f687366f0b" exitCode=0 Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.682790 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25h7x" event={"ID":"b6986f8f-6054-4fb5-bcf0-eea927f2ce77","Type":"ContainerDied","Data":"2c260652bebf9782a63a6ac5921b70138d778e4b69d3f500bdb317f687366f0b"} Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.682816 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25h7x" event={"ID":"b6986f8f-6054-4fb5-bcf0-eea927f2ce77","Type":"ContainerStarted","Data":"fde76a0b8d160aaf7f0280162b6e16ce770c92110446df60776e2a770a1af9cb"} Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.698714 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt6th\" (UniqueName: \"kubernetes.io/projected/98f8e837-97e0-48fd-a9ab-8383184a052c-kube-api-access-kt6th\") pod \"redhat-operators-sdl64\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:11 crc kubenswrapper[4946]: I1203 07:49:11.811614 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:12 crc kubenswrapper[4946]: I1203 07:49:12.011447 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdl64"] Dec 03 07:49:12 crc kubenswrapper[4946]: I1203 07:49:12.690409 4946 generic.go:334] "Generic (PLEG): container finished" podID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerID="1d38828471ea6c3937a637aca506742f58fb6693d0bd6d5060e120e3206d883f" exitCode=0 Dec 03 07:49:12 crc kubenswrapper[4946]: I1203 07:49:12.690517 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25h7x" event={"ID":"b6986f8f-6054-4fb5-bcf0-eea927f2ce77","Type":"ContainerDied","Data":"1d38828471ea6c3937a637aca506742f58fb6693d0bd6d5060e120e3206d883f"} Dec 03 07:49:12 crc kubenswrapper[4946]: I1203 07:49:12.692754 4946 generic.go:334] "Generic (PLEG): container finished" podID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerID="99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740" exitCode=0 Dec 03 07:49:12 crc kubenswrapper[4946]: I1203 07:49:12.692795 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdl64" event={"ID":"98f8e837-97e0-48fd-a9ab-8383184a052c","Type":"ContainerDied","Data":"99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740"} Dec 03 07:49:12 crc kubenswrapper[4946]: I1203 07:49:12.692855 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdl64" event={"ID":"98f8e837-97e0-48fd-a9ab-8383184a052c","Type":"ContainerStarted","Data":"eb5c31795d755b13db75b601825768c2664b22c2113e7983a7db8f34f879b5d6"} Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.707179 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdl64" event={"ID":"98f8e837-97e0-48fd-a9ab-8383184a052c","Type":"ContainerStarted","Data":"57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0"} Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.709486 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25h7x" event={"ID":"b6986f8f-6054-4fb5-bcf0-eea927f2ce77","Type":"ContainerStarted","Data":"52aa3402ab54eb5b9dbcf703ae51891402cf4cd31e997b79ba9ff96c6943bf49"} Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.763112 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-25h7x" podStartSLOduration=2.398321999 podStartE2EDuration="3.763084844s" podCreationTimestamp="2025-12-03 07:49:10 +0000 UTC" firstStartedPulling="2025-12-03 07:49:11.684659411 +0000 UTC m=+3544.481349520" lastFinishedPulling="2025-12-03 07:49:13.049422216 +0000 UTC m=+3545.846112365" observedRunningTime="2025-12-03 07:49:13.76295595 +0000 UTC m=+3546.559646070" watchObservedRunningTime="2025-12-03 07:49:13.763084844 +0000 UTC m=+3546.559774963" Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.863553 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g7z6h"] Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.864996 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.927693 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g7z6h"] Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.995409 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-catalog-content\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.995487 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrnc7\" (UniqueName: \"kubernetes.io/projected/92166526-ae74-4034-800d-2ecc14e0593b-kube-api-access-lrnc7\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:13 crc kubenswrapper[4946]: I1203 07:49:13.995566 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-utilities\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.096456 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-catalog-content\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.096509 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrnc7\" (UniqueName: \"kubernetes.io/projected/92166526-ae74-4034-800d-2ecc14e0593b-kube-api-access-lrnc7\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.096537 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-utilities\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.097095 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-catalog-content\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.097125 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-utilities\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.119970 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrnc7\" (UniqueName: \"kubernetes.io/projected/92166526-ae74-4034-800d-2ecc14e0593b-kube-api-access-lrnc7\") pod \"community-operators-g7z6h\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.194066 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.717787 4946 generic.go:334] "Generic (PLEG): container finished" podID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerID="57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0" exitCode=0 Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.717882 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdl64" event={"ID":"98f8e837-97e0-48fd-a9ab-8383184a052c","Type":"ContainerDied","Data":"57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0"} Dec 03 07:49:14 crc kubenswrapper[4946]: I1203 07:49:14.722258 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g7z6h"] Dec 03 07:49:15 crc kubenswrapper[4946]: I1203 07:49:15.725364 4946 generic.go:334] "Generic (PLEG): container finished" podID="92166526-ae74-4034-800d-2ecc14e0593b" containerID="c819f116fa8fef53a8a1de092ce4a9761c1a1a8da181723ce05aee5639e01e76" exitCode=0 Dec 03 07:49:15 crc kubenswrapper[4946]: I1203 07:49:15.725414 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7z6h" event={"ID":"92166526-ae74-4034-800d-2ecc14e0593b","Type":"ContainerDied","Data":"c819f116fa8fef53a8a1de092ce4a9761c1a1a8da181723ce05aee5639e01e76"} Dec 03 07:49:15 crc kubenswrapper[4946]: I1203 07:49:15.727642 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7z6h" event={"ID":"92166526-ae74-4034-800d-2ecc14e0593b","Type":"ContainerStarted","Data":"4b30896c065919860131a6dd722e02d829f7de41cbf7780a97cd4c6838a0b073"} Dec 03 07:49:15 crc kubenswrapper[4946]: I1203 07:49:15.730919 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdl64" event={"ID":"98f8e837-97e0-48fd-a9ab-8383184a052c","Type":"ContainerStarted","Data":"3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908"} Dec 03 07:49:15 crc kubenswrapper[4946]: I1203 07:49:15.768699 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sdl64" podStartSLOduration=2.350100761 podStartE2EDuration="4.768681334s" podCreationTimestamp="2025-12-03 07:49:11 +0000 UTC" firstStartedPulling="2025-12-03 07:49:12.695973462 +0000 UTC m=+3545.492663571" lastFinishedPulling="2025-12-03 07:49:15.114554025 +0000 UTC m=+3547.911244144" observedRunningTime="2025-12-03 07:49:15.764624095 +0000 UTC m=+3548.561314204" watchObservedRunningTime="2025-12-03 07:49:15.768681334 +0000 UTC m=+3548.565371443" Dec 03 07:49:16 crc kubenswrapper[4946]: I1203 07:49:16.739206 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7z6h" event={"ID":"92166526-ae74-4034-800d-2ecc14e0593b","Type":"ContainerStarted","Data":"77eadcf3995f5d41e1524e668cd007f93427c5019aebf5fe0b584f8db1facdb1"} Dec 03 07:49:18 crc kubenswrapper[4946]: I1203 07:49:17.753080 4946 generic.go:334] "Generic (PLEG): container finished" podID="92166526-ae74-4034-800d-2ecc14e0593b" containerID="77eadcf3995f5d41e1524e668cd007f93427c5019aebf5fe0b584f8db1facdb1" exitCode=0 Dec 03 07:49:18 crc kubenswrapper[4946]: I1203 07:49:17.753129 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7z6h" event={"ID":"92166526-ae74-4034-800d-2ecc14e0593b","Type":"ContainerDied","Data":"77eadcf3995f5d41e1524e668cd007f93427c5019aebf5fe0b584f8db1facdb1"} Dec 03 07:49:20 crc kubenswrapper[4946]: I1203 07:49:20.774334 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7z6h" event={"ID":"92166526-ae74-4034-800d-2ecc14e0593b","Type":"ContainerStarted","Data":"8729d63d79e8b7ff6404c9a29a17040c7f5a3243773fde672a5f7104db731d15"} Dec 03 07:49:20 crc kubenswrapper[4946]: I1203 07:49:20.799233 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g7z6h" podStartSLOduration=3.662572192 podStartE2EDuration="7.79920807s" podCreationTimestamp="2025-12-03 07:49:13 +0000 UTC" firstStartedPulling="2025-12-03 07:49:15.727018082 +0000 UTC m=+3548.523708191" lastFinishedPulling="2025-12-03 07:49:19.86365395 +0000 UTC m=+3552.660344069" observedRunningTime="2025-12-03 07:49:20.794846184 +0000 UTC m=+3553.591536303" watchObservedRunningTime="2025-12-03 07:49:20.79920807 +0000 UTC m=+3553.595898179" Dec 03 07:49:21 crc kubenswrapper[4946]: I1203 07:49:21.205363 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:21 crc kubenswrapper[4946]: I1203 07:49:21.206962 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:21 crc kubenswrapper[4946]: I1203 07:49:21.258785 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:21 crc kubenswrapper[4946]: I1203 07:49:21.812150 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:21 crc kubenswrapper[4946]: I1203 07:49:21.812197 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:21 crc kubenswrapper[4946]: I1203 07:49:21.830450 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:22 crc kubenswrapper[4946]: I1203 07:49:22.657804 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25h7x"] Dec 03 07:49:22 crc kubenswrapper[4946]: I1203 07:49:22.856818 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sdl64" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="registry-server" probeResult="failure" output=< Dec 03 07:49:22 crc kubenswrapper[4946]: timeout: failed to connect service ":50051" within 1s Dec 03 07:49:22 crc kubenswrapper[4946]: > Dec 03 07:49:23 crc kubenswrapper[4946]: I1203 07:49:23.804789 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-25h7x" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerName="registry-server" containerID="cri-o://52aa3402ab54eb5b9dbcf703ae51891402cf4cd31e997b79ba9ff96c6943bf49" gracePeriod=2 Dec 03 07:49:24 crc kubenswrapper[4946]: I1203 07:49:24.194255 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:24 crc kubenswrapper[4946]: I1203 07:49:24.194568 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:24 crc kubenswrapper[4946]: I1203 07:49:24.287168 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:27 crc kubenswrapper[4946]: I1203 07:49:27.840972 4946 generic.go:334] "Generic (PLEG): container finished" podID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerID="52aa3402ab54eb5b9dbcf703ae51891402cf4cd31e997b79ba9ff96c6943bf49" exitCode=0 Dec 03 07:49:27 crc kubenswrapper[4946]: I1203 07:49:27.841069 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25h7x" event={"ID":"b6986f8f-6054-4fb5-bcf0-eea927f2ce77","Type":"ContainerDied","Data":"52aa3402ab54eb5b9dbcf703ae51891402cf4cd31e997b79ba9ff96c6943bf49"} Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.319793 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.478227 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j894v\" (UniqueName: \"kubernetes.io/projected/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-kube-api-access-j894v\") pod \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.478331 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-utilities\") pod \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.478397 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-catalog-content\") pod \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\" (UID: \"b6986f8f-6054-4fb5-bcf0-eea927f2ce77\") " Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.480474 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-utilities" (OuterVolumeSpecName: "utilities") pod "b6986f8f-6054-4fb5-bcf0-eea927f2ce77" (UID: "b6986f8f-6054-4fb5-bcf0-eea927f2ce77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.485920 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-kube-api-access-j894v" (OuterVolumeSpecName: "kube-api-access-j894v") pod "b6986f8f-6054-4fb5-bcf0-eea927f2ce77" (UID: "b6986f8f-6054-4fb5-bcf0-eea927f2ce77"). InnerVolumeSpecName "kube-api-access-j894v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.508590 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6986f8f-6054-4fb5-bcf0-eea927f2ce77" (UID: "b6986f8f-6054-4fb5-bcf0-eea927f2ce77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.579637 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j894v\" (UniqueName: \"kubernetes.io/projected/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-kube-api-access-j894v\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.579673 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.579683 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6986f8f-6054-4fb5-bcf0-eea927f2ce77-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.852170 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25h7x" event={"ID":"b6986f8f-6054-4fb5-bcf0-eea927f2ce77","Type":"ContainerDied","Data":"fde76a0b8d160aaf7f0280162b6e16ce770c92110446df60776e2a770a1af9cb"} Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.852217 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25h7x" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.852239 4946 scope.go:117] "RemoveContainer" containerID="52aa3402ab54eb5b9dbcf703ae51891402cf4cd31e997b79ba9ff96c6943bf49" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.880666 4946 scope.go:117] "RemoveContainer" containerID="1d38828471ea6c3937a637aca506742f58fb6693d0bd6d5060e120e3206d883f" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.918818 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25h7x"] Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.947635 4946 scope.go:117] "RemoveContainer" containerID="2c260652bebf9782a63a6ac5921b70138d778e4b69d3f500bdb317f687366f0b" Dec 03 07:49:28 crc kubenswrapper[4946]: I1203 07:49:28.949617 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-25h7x"] Dec 03 07:49:29 crc kubenswrapper[4946]: I1203 07:49:29.606670 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" path="/var/lib/kubelet/pods/b6986f8f-6054-4fb5-bcf0-eea927f2ce77/volumes" Dec 03 07:49:31 crc kubenswrapper[4946]: I1203 07:49:31.888673 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:31 crc kubenswrapper[4946]: I1203 07:49:31.966517 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:32 crc kubenswrapper[4946]: I1203 07:49:32.140166 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdl64"] Dec 03 07:49:32 crc kubenswrapper[4946]: I1203 07:49:32.910459 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sdl64" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="registry-server" containerID="cri-o://3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908" gracePeriod=2 Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.392681 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.562347 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-catalog-content\") pod \"98f8e837-97e0-48fd-a9ab-8383184a052c\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.562627 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-utilities\") pod \"98f8e837-97e0-48fd-a9ab-8383184a052c\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.562725 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt6th\" (UniqueName: \"kubernetes.io/projected/98f8e837-97e0-48fd-a9ab-8383184a052c-kube-api-access-kt6th\") pod \"98f8e837-97e0-48fd-a9ab-8383184a052c\" (UID: \"98f8e837-97e0-48fd-a9ab-8383184a052c\") " Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.564039 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-utilities" (OuterVolumeSpecName: "utilities") pod "98f8e837-97e0-48fd-a9ab-8383184a052c" (UID: "98f8e837-97e0-48fd-a9ab-8383184a052c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.564424 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.571980 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98f8e837-97e0-48fd-a9ab-8383184a052c-kube-api-access-kt6th" (OuterVolumeSpecName: "kube-api-access-kt6th") pod "98f8e837-97e0-48fd-a9ab-8383184a052c" (UID: "98f8e837-97e0-48fd-a9ab-8383184a052c"). InnerVolumeSpecName "kube-api-access-kt6th". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.666176 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt6th\" (UniqueName: \"kubernetes.io/projected/98f8e837-97e0-48fd-a9ab-8383184a052c-kube-api-access-kt6th\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.734162 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98f8e837-97e0-48fd-a9ab-8383184a052c" (UID: "98f8e837-97e0-48fd-a9ab-8383184a052c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.767844 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f8e837-97e0-48fd-a9ab-8383184a052c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.936457 4946 generic.go:334] "Generic (PLEG): container finished" podID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerID="3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908" exitCode=0 Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.936515 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdl64" event={"ID":"98f8e837-97e0-48fd-a9ab-8383184a052c","Type":"ContainerDied","Data":"3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908"} Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.936542 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdl64" event={"ID":"98f8e837-97e0-48fd-a9ab-8383184a052c","Type":"ContainerDied","Data":"eb5c31795d755b13db75b601825768c2664b22c2113e7983a7db8f34f879b5d6"} Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.936568 4946 scope.go:117] "RemoveContainer" containerID="3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.936756 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdl64" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.979005 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdl64"] Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.979385 4946 scope.go:117] "RemoveContainer" containerID="57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0" Dec 03 07:49:33 crc kubenswrapper[4946]: I1203 07:49:33.985280 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sdl64"] Dec 03 07:49:34 crc kubenswrapper[4946]: I1203 07:49:34.014628 4946 scope.go:117] "RemoveContainer" containerID="99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740" Dec 03 07:49:34 crc kubenswrapper[4946]: I1203 07:49:34.040047 4946 scope.go:117] "RemoveContainer" containerID="3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908" Dec 03 07:49:34 crc kubenswrapper[4946]: E1203 07:49:34.040548 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908\": container with ID starting with 3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908 not found: ID does not exist" containerID="3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908" Dec 03 07:49:34 crc kubenswrapper[4946]: I1203 07:49:34.040584 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908"} err="failed to get container status \"3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908\": rpc error: code = NotFound desc = could not find container \"3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908\": container with ID starting with 3af4be59d151650b0f2b43bd3a460c678ab0db27c0f33d015f7f3e52e8eef908 not found: ID does not exist" Dec 03 07:49:34 crc kubenswrapper[4946]: I1203 07:49:34.040610 4946 scope.go:117] "RemoveContainer" containerID="57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0" Dec 03 07:49:34 crc kubenswrapper[4946]: E1203 07:49:34.041313 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0\": container with ID starting with 57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0 not found: ID does not exist" containerID="57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0" Dec 03 07:49:34 crc kubenswrapper[4946]: I1203 07:49:34.041346 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0"} err="failed to get container status \"57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0\": rpc error: code = NotFound desc = could not find container \"57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0\": container with ID starting with 57f620c710be7f700d66d32f4007be25d9d2223aa0dc011e8bda3bad6f19cdd0 not found: ID does not exist" Dec 03 07:49:34 crc kubenswrapper[4946]: I1203 07:49:34.041364 4946 scope.go:117] "RemoveContainer" containerID="99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740" Dec 03 07:49:34 crc kubenswrapper[4946]: E1203 07:49:34.042791 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740\": container with ID starting with 99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740 not found: ID does not exist" containerID="99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740" Dec 03 07:49:34 crc kubenswrapper[4946]: I1203 07:49:34.042874 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740"} err="failed to get container status \"99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740\": rpc error: code = NotFound desc = could not find container \"99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740\": container with ID starting with 99b2337bd12d63b23de5af82c248a2dc1c5c2e43fffe2c296ada2fc42c6f7740 not found: ID does not exist" Dec 03 07:49:34 crc kubenswrapper[4946]: I1203 07:49:34.267681 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:35 crc kubenswrapper[4946]: I1203 07:49:35.606147 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" path="/var/lib/kubelet/pods/98f8e837-97e0-48fd-a9ab-8383184a052c/volumes" Dec 03 07:49:36 crc kubenswrapper[4946]: I1203 07:49:36.536767 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g7z6h"] Dec 03 07:49:36 crc kubenswrapper[4946]: I1203 07:49:36.537115 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g7z6h" podUID="92166526-ae74-4034-800d-2ecc14e0593b" containerName="registry-server" containerID="cri-o://8729d63d79e8b7ff6404c9a29a17040c7f5a3243773fde672a5f7104db731d15" gracePeriod=2 Dec 03 07:49:36 crc kubenswrapper[4946]: I1203 07:49:36.965519 4946 generic.go:334] "Generic (PLEG): container finished" podID="92166526-ae74-4034-800d-2ecc14e0593b" containerID="8729d63d79e8b7ff6404c9a29a17040c7f5a3243773fde672a5f7104db731d15" exitCode=0 Dec 03 07:49:36 crc kubenswrapper[4946]: I1203 07:49:36.965636 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7z6h" event={"ID":"92166526-ae74-4034-800d-2ecc14e0593b","Type":"ContainerDied","Data":"8729d63d79e8b7ff6404c9a29a17040c7f5a3243773fde672a5f7104db731d15"} Dec 03 07:49:36 crc kubenswrapper[4946]: I1203 07:49:36.966101 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7z6h" event={"ID":"92166526-ae74-4034-800d-2ecc14e0593b","Type":"ContainerDied","Data":"4b30896c065919860131a6dd722e02d829f7de41cbf7780a97cd4c6838a0b073"} Dec 03 07:49:36 crc kubenswrapper[4946]: I1203 07:49:36.966127 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b30896c065919860131a6dd722e02d829f7de41cbf7780a97cd4c6838a0b073" Dec 03 07:49:36 crc kubenswrapper[4946]: I1203 07:49:36.973853 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.022971 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrnc7\" (UniqueName: \"kubernetes.io/projected/92166526-ae74-4034-800d-2ecc14e0593b-kube-api-access-lrnc7\") pod \"92166526-ae74-4034-800d-2ecc14e0593b\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.023037 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-utilities\") pod \"92166526-ae74-4034-800d-2ecc14e0593b\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.023075 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-catalog-content\") pod \"92166526-ae74-4034-800d-2ecc14e0593b\" (UID: \"92166526-ae74-4034-800d-2ecc14e0593b\") " Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.025196 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-utilities" (OuterVolumeSpecName: "utilities") pod "92166526-ae74-4034-800d-2ecc14e0593b" (UID: "92166526-ae74-4034-800d-2ecc14e0593b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.036268 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92166526-ae74-4034-800d-2ecc14e0593b-kube-api-access-lrnc7" (OuterVolumeSpecName: "kube-api-access-lrnc7") pod "92166526-ae74-4034-800d-2ecc14e0593b" (UID: "92166526-ae74-4034-800d-2ecc14e0593b"). InnerVolumeSpecName "kube-api-access-lrnc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.076965 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92166526-ae74-4034-800d-2ecc14e0593b" (UID: "92166526-ae74-4034-800d-2ecc14e0593b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.124988 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrnc7\" (UniqueName: \"kubernetes.io/projected/92166526-ae74-4034-800d-2ecc14e0593b-kube-api-access-lrnc7\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.125038 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.125060 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92166526-ae74-4034-800d-2ecc14e0593b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:49:37 crc kubenswrapper[4946]: I1203 07:49:37.975981 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g7z6h" Dec 03 07:49:38 crc kubenswrapper[4946]: I1203 07:49:38.018910 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g7z6h"] Dec 03 07:49:38 crc kubenswrapper[4946]: I1203 07:49:38.029486 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g7z6h"] Dec 03 07:49:39 crc kubenswrapper[4946]: I1203 07:49:39.602966 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92166526-ae74-4034-800d-2ecc14e0593b" path="/var/lib/kubelet/pods/92166526-ae74-4034-800d-2ecc14e0593b/volumes" Dec 03 07:50:23 crc kubenswrapper[4946]: I1203 07:50:23.039404 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:50:23 crc kubenswrapper[4946]: I1203 07:50:23.040244 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:50:53 crc kubenswrapper[4946]: I1203 07:50:53.039131 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:50:53 crc kubenswrapper[4946]: I1203 07:50:53.039793 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.039460 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.040141 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.040185 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.040922 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.040991 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" gracePeriod=600 Dec 03 07:51:23 crc kubenswrapper[4946]: E1203 07:51:23.188320 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.943648 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" exitCode=0 Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.943708 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9"} Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.943759 4946 scope.go:117] "RemoveContainer" containerID="ae1d7773e3ac40586172ec68cf3a08b2f4a5181ad947298f33fcaa0b2cb770dc" Dec 03 07:51:23 crc kubenswrapper[4946]: I1203 07:51:23.944322 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:51:23 crc kubenswrapper[4946]: E1203 07:51:23.944602 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:51:34 crc kubenswrapper[4946]: I1203 07:51:34.592879 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:51:34 crc kubenswrapper[4946]: E1203 07:51:34.593868 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:51:47 crc kubenswrapper[4946]: I1203 07:51:47.601779 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:51:47 crc kubenswrapper[4946]: E1203 07:51:47.602650 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:51:58 crc kubenswrapper[4946]: I1203 07:51:58.593201 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:51:58 crc kubenswrapper[4946]: E1203 07:51:58.593975 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:52:09 crc kubenswrapper[4946]: I1203 07:52:09.592815 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:52:09 crc kubenswrapper[4946]: E1203 07:52:09.594147 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:52:21 crc kubenswrapper[4946]: I1203 07:52:21.593813 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:52:21 crc kubenswrapper[4946]: E1203 07:52:21.594830 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:52:34 crc kubenswrapper[4946]: I1203 07:52:34.593026 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:52:34 crc kubenswrapper[4946]: E1203 07:52:34.595247 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:52:45 crc kubenswrapper[4946]: I1203 07:52:45.593590 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:52:45 crc kubenswrapper[4946]: E1203 07:52:45.594595 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:52:58 crc kubenswrapper[4946]: I1203 07:52:58.594667 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:52:58 crc kubenswrapper[4946]: E1203 07:52:58.597080 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:53:10 crc kubenswrapper[4946]: I1203 07:53:10.593032 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:53:10 crc kubenswrapper[4946]: E1203 07:53:10.594301 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:53:23 crc kubenswrapper[4946]: I1203 07:53:23.593896 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:53:23 crc kubenswrapper[4946]: E1203 07:53:23.595072 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:53:34 crc kubenswrapper[4946]: I1203 07:53:34.592989 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:53:34 crc kubenswrapper[4946]: E1203 07:53:34.594164 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:53:47 crc kubenswrapper[4946]: I1203 07:53:47.607428 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:53:47 crc kubenswrapper[4946]: E1203 07:53:47.608851 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:54:01 crc kubenswrapper[4946]: I1203 07:54:01.592940 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:54:01 crc kubenswrapper[4946]: E1203 07:54:01.594084 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:54:16 crc kubenswrapper[4946]: I1203 07:54:16.593162 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:54:16 crc kubenswrapper[4946]: E1203 07:54:16.594338 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:54:29 crc kubenswrapper[4946]: I1203 07:54:29.593266 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:54:29 crc kubenswrapper[4946]: E1203 07:54:29.594265 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:54:41 crc kubenswrapper[4946]: I1203 07:54:41.593309 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:54:41 crc kubenswrapper[4946]: E1203 07:54:41.593945 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:54:52 crc kubenswrapper[4946]: I1203 07:54:52.593061 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:54:52 crc kubenswrapper[4946]: E1203 07:54:52.594070 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:55:05 crc kubenswrapper[4946]: I1203 07:55:05.593887 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:55:05 crc kubenswrapper[4946]: E1203 07:55:05.594957 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:55:19 crc kubenswrapper[4946]: I1203 07:55:19.592411 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:55:19 crc kubenswrapper[4946]: E1203 07:55:19.593022 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:55:31 crc kubenswrapper[4946]: I1203 07:55:31.593226 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:55:31 crc kubenswrapper[4946]: E1203 07:55:31.594405 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:55:42 crc kubenswrapper[4946]: I1203 07:55:42.593073 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:55:42 crc kubenswrapper[4946]: E1203 07:55:42.594111 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:55:55 crc kubenswrapper[4946]: I1203 07:55:55.592562 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:55:55 crc kubenswrapper[4946]: E1203 07:55:55.593303 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:56:06 crc kubenswrapper[4946]: I1203 07:56:06.593255 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:56:06 crc kubenswrapper[4946]: E1203 07:56:06.594315 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:56:13 crc kubenswrapper[4946]: I1203 07:56:13.698927 4946 scope.go:117] "RemoveContainer" containerID="8729d63d79e8b7ff6404c9a29a17040c7f5a3243773fde672a5f7104db731d15" Dec 03 07:56:13 crc kubenswrapper[4946]: I1203 07:56:13.729761 4946 scope.go:117] "RemoveContainer" containerID="c819f116fa8fef53a8a1de092ce4a9761c1a1a8da181723ce05aee5639e01e76" Dec 03 07:56:13 crc kubenswrapper[4946]: I1203 07:56:13.754570 4946 scope.go:117] "RemoveContainer" containerID="77eadcf3995f5d41e1524e668cd007f93427c5019aebf5fe0b584f8db1facdb1" Dec 03 07:56:18 crc kubenswrapper[4946]: I1203 07:56:18.592731 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:56:18 crc kubenswrapper[4946]: E1203 07:56:18.593422 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 07:56:31 crc kubenswrapper[4946]: I1203 07:56:31.593799 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:56:32 crc kubenswrapper[4946]: I1203 07:56:32.747213 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"ca923913b5c35f3d0179d87c55d99b6dc0fcac16a63f3f56c27146bef1002661"} Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.598880 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hsdv2"] Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600330 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="extract-utilities" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600363 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="extract-utilities" Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600387 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92166526-ae74-4034-800d-2ecc14e0593b" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600399 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="92166526-ae74-4034-800d-2ecc14e0593b" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600419 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="extract-content" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600433 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="extract-content" Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600461 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerName="extract-utilities" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600473 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerName="extract-utilities" Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600499 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92166526-ae74-4034-800d-2ecc14e0593b" containerName="extract-utilities" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600513 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="92166526-ae74-4034-800d-2ecc14e0593b" containerName="extract-utilities" Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600542 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600554 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600582 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92166526-ae74-4034-800d-2ecc14e0593b" containerName="extract-content" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600594 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="92166526-ae74-4034-800d-2ecc14e0593b" containerName="extract-content" Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600611 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerName="extract-content" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600623 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerName="extract-content" Dec 03 07:57:50 crc kubenswrapper[4946]: E1203 07:57:50.600644 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600655 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600938 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="92166526-ae74-4034-800d-2ecc14e0593b" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600970 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="98f8e837-97e0-48fd-a9ab-8383184a052c" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.600995 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6986f8f-6054-4fb5-bcf0-eea927f2ce77" containerName="registry-server" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.603023 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.616538 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsdv2"] Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.728354 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctpzw\" (UniqueName: \"kubernetes.io/projected/e5e5669e-63e7-4bae-892a-04d994f9b400-kube-api-access-ctpzw\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.728424 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5e5669e-63e7-4bae-892a-04d994f9b400-catalog-content\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.728530 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5e5669e-63e7-4bae-892a-04d994f9b400-utilities\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.830207 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5e5669e-63e7-4bae-892a-04d994f9b400-utilities\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.830303 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctpzw\" (UniqueName: \"kubernetes.io/projected/e5e5669e-63e7-4bae-892a-04d994f9b400-kube-api-access-ctpzw\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.830331 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5e5669e-63e7-4bae-892a-04d994f9b400-catalog-content\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.830857 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5e5669e-63e7-4bae-892a-04d994f9b400-utilities\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.830905 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5e5669e-63e7-4bae-892a-04d994f9b400-catalog-content\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.849141 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctpzw\" (UniqueName: \"kubernetes.io/projected/e5e5669e-63e7-4bae-892a-04d994f9b400-kube-api-access-ctpzw\") pod \"certified-operators-hsdv2\" (UID: \"e5e5669e-63e7-4bae-892a-04d994f9b400\") " pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:50 crc kubenswrapper[4946]: I1203 07:57:50.958236 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:57:51 crc kubenswrapper[4946]: I1203 07:57:51.415632 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsdv2"] Dec 03 07:57:51 crc kubenswrapper[4946]: I1203 07:57:51.492204 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdv2" event={"ID":"e5e5669e-63e7-4bae-892a-04d994f9b400","Type":"ContainerStarted","Data":"f02db0d31ab034e0df4bef09cd0c31b75d9526f60a62aa93a55ae0fd4372dd21"} Dec 03 07:57:52 crc kubenswrapper[4946]: I1203 07:57:52.504805 4946 generic.go:334] "Generic (PLEG): container finished" podID="e5e5669e-63e7-4bae-892a-04d994f9b400" containerID="438bf05e00fddd19e6d79807c247ae6b7461e584df1080d1fc0bf0605740c920" exitCode=0 Dec 03 07:57:52 crc kubenswrapper[4946]: I1203 07:57:52.504883 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdv2" event={"ID":"e5e5669e-63e7-4bae-892a-04d994f9b400","Type":"ContainerDied","Data":"438bf05e00fddd19e6d79807c247ae6b7461e584df1080d1fc0bf0605740c920"} Dec 03 07:57:52 crc kubenswrapper[4946]: I1203 07:57:52.508083 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:57:56 crc kubenswrapper[4946]: I1203 07:57:56.540903 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdv2" event={"ID":"e5e5669e-63e7-4bae-892a-04d994f9b400","Type":"ContainerStarted","Data":"1bdf79ba97f77a4262ecf2d4709065f40fb5ba3e48a055e80837357870c97e4e"} Dec 03 07:57:57 crc kubenswrapper[4946]: I1203 07:57:57.551571 4946 generic.go:334] "Generic (PLEG): container finished" podID="e5e5669e-63e7-4bae-892a-04d994f9b400" containerID="1bdf79ba97f77a4262ecf2d4709065f40fb5ba3e48a055e80837357870c97e4e" exitCode=0 Dec 03 07:57:57 crc kubenswrapper[4946]: I1203 07:57:57.551698 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdv2" event={"ID":"e5e5669e-63e7-4bae-892a-04d994f9b400","Type":"ContainerDied","Data":"1bdf79ba97f77a4262ecf2d4709065f40fb5ba3e48a055e80837357870c97e4e"} Dec 03 07:57:57 crc kubenswrapper[4946]: I1203 07:57:57.552330 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdv2" event={"ID":"e5e5669e-63e7-4bae-892a-04d994f9b400","Type":"ContainerStarted","Data":"f15a3faa9f94448bb5ec4e38a1b4a38e25c8905ccca6819addd589324342cdad"} Dec 03 07:57:57 crc kubenswrapper[4946]: I1203 07:57:57.586845 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hsdv2" podStartSLOduration=3.0974521250000002 podStartE2EDuration="7.586819827s" podCreationTimestamp="2025-12-03 07:57:50 +0000 UTC" firstStartedPulling="2025-12-03 07:57:52.507508039 +0000 UTC m=+4065.304198158" lastFinishedPulling="2025-12-03 07:57:56.996875751 +0000 UTC m=+4069.793565860" observedRunningTime="2025-12-03 07:57:57.578131355 +0000 UTC m=+4070.374821504" watchObservedRunningTime="2025-12-03 07:57:57.586819827 +0000 UTC m=+4070.383509966" Dec 03 07:58:00 crc kubenswrapper[4946]: I1203 07:58:00.959257 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:58:00 crc kubenswrapper[4946]: I1203 07:58:00.959711 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:58:01 crc kubenswrapper[4946]: I1203 07:58:01.025054 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.030619 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hsdv2" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.115055 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsdv2"] Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.162423 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-29nz7"] Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.162693 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-29nz7" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerName="registry-server" containerID="cri-o://040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5" gracePeriod=2 Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.614503 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-29nz7" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.678330 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c59tl\" (UniqueName: \"kubernetes.io/projected/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-kube-api-access-c59tl\") pod \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.678400 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-catalog-content\") pod \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.678432 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-utilities\") pod \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\" (UID: \"b1f9a89f-9eef-475c-8d1c-3fcf430187cf\") " Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.679196 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-utilities" (OuterVolumeSpecName: "utilities") pod "b1f9a89f-9eef-475c-8d1c-3fcf430187cf" (UID: "b1f9a89f-9eef-475c-8d1c-3fcf430187cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.684766 4946 generic.go:334] "Generic (PLEG): container finished" podID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerID="040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5" exitCode=0 Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.685682 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-29nz7" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.686014 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nz7" event={"ID":"b1f9a89f-9eef-475c-8d1c-3fcf430187cf","Type":"ContainerDied","Data":"040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5"} Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.686044 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nz7" event={"ID":"b1f9a89f-9eef-475c-8d1c-3fcf430187cf","Type":"ContainerDied","Data":"906e3f55961b7dd2b0b277fc40efba13dd7b7dce2081c41e50a184031db6b10a"} Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.686060 4946 scope.go:117] "RemoveContainer" containerID="040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.699955 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-kube-api-access-c59tl" (OuterVolumeSpecName: "kube-api-access-c59tl") pod "b1f9a89f-9eef-475c-8d1c-3fcf430187cf" (UID: "b1f9a89f-9eef-475c-8d1c-3fcf430187cf"). InnerVolumeSpecName "kube-api-access-c59tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.727432 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1f9a89f-9eef-475c-8d1c-3fcf430187cf" (UID: "b1f9a89f-9eef-475c-8d1c-3fcf430187cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.730069 4946 scope.go:117] "RemoveContainer" containerID="422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.748819 4946 scope.go:117] "RemoveContainer" containerID="7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.771153 4946 scope.go:117] "RemoveContainer" containerID="040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5" Dec 03 07:58:11 crc kubenswrapper[4946]: E1203 07:58:11.771652 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5\": container with ID starting with 040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5 not found: ID does not exist" containerID="040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.771693 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5"} err="failed to get container status \"040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5\": rpc error: code = NotFound desc = could not find container \"040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5\": container with ID starting with 040c2295c113cac779ab8a7e222bc2a6f58bb4ef51e9f6ebf3aea8509484f3d5 not found: ID does not exist" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.771822 4946 scope.go:117] "RemoveContainer" containerID="422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6" Dec 03 07:58:11 crc kubenswrapper[4946]: E1203 07:58:11.772080 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6\": container with ID starting with 422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6 not found: ID does not exist" containerID="422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.772115 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6"} err="failed to get container status \"422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6\": rpc error: code = NotFound desc = could not find container \"422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6\": container with ID starting with 422de986bee0f9e8f4b448ee5018f041f25a1f219d08086fda69620937357bb6 not found: ID does not exist" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.772133 4946 scope.go:117] "RemoveContainer" containerID="7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187" Dec 03 07:58:11 crc kubenswrapper[4946]: E1203 07:58:11.772349 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187\": container with ID starting with 7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187 not found: ID does not exist" containerID="7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.772378 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187"} err="failed to get container status \"7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187\": rpc error: code = NotFound desc = could not find container \"7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187\": container with ID starting with 7876bec426f513893fbd67644aea5db88c6f3498b3649e9c5422f575fe029187 not found: ID does not exist" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.780330 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c59tl\" (UniqueName: \"kubernetes.io/projected/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-kube-api-access-c59tl\") on node \"crc\" DevicePath \"\"" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.780369 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:58:11 crc kubenswrapper[4946]: I1203 07:58:11.780382 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f9a89f-9eef-475c-8d1c-3fcf430187cf-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:58:12 crc kubenswrapper[4946]: I1203 07:58:12.023541 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-29nz7"] Dec 03 07:58:12 crc kubenswrapper[4946]: I1203 07:58:12.031895 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-29nz7"] Dec 03 07:58:13 crc kubenswrapper[4946]: I1203 07:58:13.601134 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" path="/var/lib/kubelet/pods/b1f9a89f-9eef-475c-8d1c-3fcf430187cf/volumes" Dec 03 07:58:53 crc kubenswrapper[4946]: I1203 07:58:53.039545 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:58:53 crc kubenswrapper[4946]: I1203 07:58:53.040289 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:59:23 crc kubenswrapper[4946]: I1203 07:59:23.040002 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:59:23 crc kubenswrapper[4946]: I1203 07:59:23.040516 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:59:30 crc kubenswrapper[4946]: I1203 07:59:30.944633 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rwszw"] Dec 03 07:59:30 crc kubenswrapper[4946]: E1203 07:59:30.948317 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerName="extract-content" Dec 03 07:59:30 crc kubenswrapper[4946]: I1203 07:59:30.948543 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerName="extract-content" Dec 03 07:59:30 crc kubenswrapper[4946]: E1203 07:59:30.948973 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerName="extract-utilities" Dec 03 07:59:30 crc kubenswrapper[4946]: I1203 07:59:30.949191 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerName="extract-utilities" Dec 03 07:59:30 crc kubenswrapper[4946]: E1203 07:59:30.949400 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerName="registry-server" Dec 03 07:59:30 crc kubenswrapper[4946]: I1203 07:59:30.949578 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerName="registry-server" Dec 03 07:59:30 crc kubenswrapper[4946]: I1203 07:59:30.950254 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1f9a89f-9eef-475c-8d1c-3fcf430187cf" containerName="registry-server" Dec 03 07:59:30 crc kubenswrapper[4946]: I1203 07:59:30.953187 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:30 crc kubenswrapper[4946]: I1203 07:59:30.984561 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rwszw"] Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.107577 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-catalog-content\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.107673 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-utilities\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.107904 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-642bx\" (UniqueName: \"kubernetes.io/projected/f2e1a674-4872-439d-ab90-acf234b8d250-kube-api-access-642bx\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.209065 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-642bx\" (UniqueName: \"kubernetes.io/projected/f2e1a674-4872-439d-ab90-acf234b8d250-kube-api-access-642bx\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.209196 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-catalog-content\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.209237 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-utilities\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.209919 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-utilities\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.210024 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-catalog-content\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.234988 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-642bx\" (UniqueName: \"kubernetes.io/projected/f2e1a674-4872-439d-ab90-acf234b8d250-kube-api-access-642bx\") pod \"redhat-operators-rwszw\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.313459 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:31 crc kubenswrapper[4946]: I1203 07:59:31.809448 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rwszw"] Dec 03 07:59:31 crc kubenswrapper[4946]: W1203 07:59:31.822220 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2e1a674_4872_439d_ab90_acf234b8d250.slice/crio-765efebd44ed80fcf8c240bb168af0e40ac90c9bc37b2ea5c4b3409ed6065163 WatchSource:0}: Error finding container 765efebd44ed80fcf8c240bb168af0e40ac90c9bc37b2ea5c4b3409ed6065163: Status 404 returned error can't find the container with id 765efebd44ed80fcf8c240bb168af0e40ac90c9bc37b2ea5c4b3409ed6065163 Dec 03 07:59:32 crc kubenswrapper[4946]: I1203 07:59:32.465886 4946 generic.go:334] "Generic (PLEG): container finished" podID="f2e1a674-4872-439d-ab90-acf234b8d250" containerID="b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca" exitCode=0 Dec 03 07:59:32 crc kubenswrapper[4946]: I1203 07:59:32.465998 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rwszw" event={"ID":"f2e1a674-4872-439d-ab90-acf234b8d250","Type":"ContainerDied","Data":"b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca"} Dec 03 07:59:32 crc kubenswrapper[4946]: I1203 07:59:32.466310 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rwszw" event={"ID":"f2e1a674-4872-439d-ab90-acf234b8d250","Type":"ContainerStarted","Data":"765efebd44ed80fcf8c240bb168af0e40ac90c9bc37b2ea5c4b3409ed6065163"} Dec 03 07:59:33 crc kubenswrapper[4946]: I1203 07:59:33.477183 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rwszw" event={"ID":"f2e1a674-4872-439d-ab90-acf234b8d250","Type":"ContainerStarted","Data":"58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e"} Dec 03 07:59:34 crc kubenswrapper[4946]: I1203 07:59:34.486759 4946 generic.go:334] "Generic (PLEG): container finished" podID="f2e1a674-4872-439d-ab90-acf234b8d250" containerID="58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e" exitCode=0 Dec 03 07:59:34 crc kubenswrapper[4946]: I1203 07:59:34.486855 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rwszw" event={"ID":"f2e1a674-4872-439d-ab90-acf234b8d250","Type":"ContainerDied","Data":"58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e"} Dec 03 07:59:36 crc kubenswrapper[4946]: I1203 07:59:36.509357 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rwszw" event={"ID":"f2e1a674-4872-439d-ab90-acf234b8d250","Type":"ContainerStarted","Data":"e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca"} Dec 03 07:59:36 crc kubenswrapper[4946]: I1203 07:59:36.536660 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rwszw" podStartSLOduration=3.78871907 podStartE2EDuration="6.536629373s" podCreationTimestamp="2025-12-03 07:59:30 +0000 UTC" firstStartedPulling="2025-12-03 07:59:32.46914891 +0000 UTC m=+4165.265839049" lastFinishedPulling="2025-12-03 07:59:35.217059243 +0000 UTC m=+4168.013749352" observedRunningTime="2025-12-03 07:59:36.527513369 +0000 UTC m=+4169.324203488" watchObservedRunningTime="2025-12-03 07:59:36.536629373 +0000 UTC m=+4169.333319522" Dec 03 07:59:41 crc kubenswrapper[4946]: I1203 07:59:41.314530 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:41 crc kubenswrapper[4946]: I1203 07:59:41.315259 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:42 crc kubenswrapper[4946]: I1203 07:59:42.372088 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rwszw" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="registry-server" probeResult="failure" output=< Dec 03 07:59:42 crc kubenswrapper[4946]: timeout: failed to connect service ":50051" within 1s Dec 03 07:59:42 crc kubenswrapper[4946]: > Dec 03 07:59:51 crc kubenswrapper[4946]: I1203 07:59:51.367600 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:51 crc kubenswrapper[4946]: I1203 07:59:51.439433 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:51 crc kubenswrapper[4946]: I1203 07:59:51.616006 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rwszw"] Dec 03 07:59:52 crc kubenswrapper[4946]: I1203 07:59:52.650516 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rwszw" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="registry-server" containerID="cri-o://e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca" gracePeriod=2 Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.039273 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.039328 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.039369 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.040022 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca923913b5c35f3d0179d87c55d99b6dc0fcac16a63f3f56c27146bef1002661"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.040093 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://ca923913b5c35f3d0179d87c55d99b6dc0fcac16a63f3f56c27146bef1002661" gracePeriod=600 Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.293507 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.403258 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-catalog-content\") pod \"f2e1a674-4872-439d-ab90-acf234b8d250\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.403407 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-utilities\") pod \"f2e1a674-4872-439d-ab90-acf234b8d250\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.403452 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-642bx\" (UniqueName: \"kubernetes.io/projected/f2e1a674-4872-439d-ab90-acf234b8d250-kube-api-access-642bx\") pod \"f2e1a674-4872-439d-ab90-acf234b8d250\" (UID: \"f2e1a674-4872-439d-ab90-acf234b8d250\") " Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.405329 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-utilities" (OuterVolumeSpecName: "utilities") pod "f2e1a674-4872-439d-ab90-acf234b8d250" (UID: "f2e1a674-4872-439d-ab90-acf234b8d250"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.417042 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2e1a674-4872-439d-ab90-acf234b8d250-kube-api-access-642bx" (OuterVolumeSpecName: "kube-api-access-642bx") pod "f2e1a674-4872-439d-ab90-acf234b8d250" (UID: "f2e1a674-4872-439d-ab90-acf234b8d250"). InnerVolumeSpecName "kube-api-access-642bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.505080 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.505126 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-642bx\" (UniqueName: \"kubernetes.io/projected/f2e1a674-4872-439d-ab90-acf234b8d250-kube-api-access-642bx\") on node \"crc\" DevicePath \"\"" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.523658 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2e1a674-4872-439d-ab90-acf234b8d250" (UID: "f2e1a674-4872-439d-ab90-acf234b8d250"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.606230 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2e1a674-4872-439d-ab90-acf234b8d250-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.659795 4946 generic.go:334] "Generic (PLEG): container finished" podID="f2e1a674-4872-439d-ab90-acf234b8d250" containerID="e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca" exitCode=0 Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.659872 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rwszw" event={"ID":"f2e1a674-4872-439d-ab90-acf234b8d250","Type":"ContainerDied","Data":"e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca"} Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.659905 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rwszw" event={"ID":"f2e1a674-4872-439d-ab90-acf234b8d250","Type":"ContainerDied","Data":"765efebd44ed80fcf8c240bb168af0e40ac90c9bc37b2ea5c4b3409ed6065163"} Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.659899 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rwszw" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.659943 4946 scope.go:117] "RemoveContainer" containerID="e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.666302 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="ca923913b5c35f3d0179d87c55d99b6dc0fcac16a63f3f56c27146bef1002661" exitCode=0 Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.666329 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"ca923913b5c35f3d0179d87c55d99b6dc0fcac16a63f3f56c27146bef1002661"} Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.666509 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f"} Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.689240 4946 scope.go:117] "RemoveContainer" containerID="58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.711623 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rwszw"] Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.721430 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rwszw"] Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.723722 4946 scope.go:117] "RemoveContainer" containerID="b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.748150 4946 scope.go:117] "RemoveContainer" containerID="e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca" Dec 03 07:59:53 crc kubenswrapper[4946]: E1203 07:59:53.748719 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca\": container with ID starting with e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca not found: ID does not exist" containerID="e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.748802 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca"} err="failed to get container status \"e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca\": rpc error: code = NotFound desc = could not find container \"e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca\": container with ID starting with e98b9a873c9756b3fda6b2e2eb9709c0eb27772c7a26eb961bb7418a8c7486ca not found: ID does not exist" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.748844 4946 scope.go:117] "RemoveContainer" containerID="58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e" Dec 03 07:59:53 crc kubenswrapper[4946]: E1203 07:59:53.749205 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e\": container with ID starting with 58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e not found: ID does not exist" containerID="58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.749258 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e"} err="failed to get container status \"58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e\": rpc error: code = NotFound desc = could not find container \"58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e\": container with ID starting with 58235b90cd9adf52e81bab25d36b5008246ba4d8b687998f545bbe30cc9a278e not found: ID does not exist" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.749292 4946 scope.go:117] "RemoveContainer" containerID="b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca" Dec 03 07:59:53 crc kubenswrapper[4946]: E1203 07:59:53.749793 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca\": container with ID starting with b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca not found: ID does not exist" containerID="b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.749841 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca"} err="failed to get container status \"b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca\": rpc error: code = NotFound desc = could not find container \"b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca\": container with ID starting with b5dc3ab01cf21c61ee2ed9adbcfd4cb9d4a1940436fa3cda3f91eb3c57b0d4ca not found: ID does not exist" Dec 03 07:59:53 crc kubenswrapper[4946]: I1203 07:59:53.749908 4946 scope.go:117] "RemoveContainer" containerID="1cd3fd69999a1ae0c29105289f274a5e0839f29013590de2aa187138c0d363d9" Dec 03 07:59:55 crc kubenswrapper[4946]: I1203 07:59:55.604712 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" path="/var/lib/kubelet/pods/f2e1a674-4872-439d-ab90-acf234b8d250/volumes" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.193525 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq"] Dec 03 08:00:00 crc kubenswrapper[4946]: E1203 08:00:00.194320 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="extract-content" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.194332 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="extract-content" Dec 03 08:00:00 crc kubenswrapper[4946]: E1203 08:00:00.194346 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="extract-utilities" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.194353 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="extract-utilities" Dec 03 08:00:00 crc kubenswrapper[4946]: E1203 08:00:00.194365 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="registry-server" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.194371 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="registry-server" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.194519 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2e1a674-4872-439d-ab90-acf234b8d250" containerName="registry-server" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.194987 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.197565 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.197777 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.209557 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq"] Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.228688 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd89e07-ac18-45a6-9713-06cfc08bf40a-config-volume\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.228835 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs7lr\" (UniqueName: \"kubernetes.io/projected/bbd89e07-ac18-45a6-9713-06cfc08bf40a-kube-api-access-xs7lr\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.228884 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd89e07-ac18-45a6-9713-06cfc08bf40a-secret-volume\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.330645 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs7lr\" (UniqueName: \"kubernetes.io/projected/bbd89e07-ac18-45a6-9713-06cfc08bf40a-kube-api-access-xs7lr\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.330722 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd89e07-ac18-45a6-9713-06cfc08bf40a-secret-volume\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.330813 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd89e07-ac18-45a6-9713-06cfc08bf40a-config-volume\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.332117 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd89e07-ac18-45a6-9713-06cfc08bf40a-config-volume\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.340079 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd89e07-ac18-45a6-9713-06cfc08bf40a-secret-volume\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.349162 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs7lr\" (UniqueName: \"kubernetes.io/projected/bbd89e07-ac18-45a6-9713-06cfc08bf40a-kube-api-access-xs7lr\") pod \"collect-profiles-29412480-vlkgq\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:00 crc kubenswrapper[4946]: I1203 08:00:00.544214 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:01 crc kubenswrapper[4946]: I1203 08:00:01.018046 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq"] Dec 03 08:00:01 crc kubenswrapper[4946]: I1203 08:00:01.737528 4946 generic.go:334] "Generic (PLEG): container finished" podID="bbd89e07-ac18-45a6-9713-06cfc08bf40a" containerID="779cad687350dfef33e55991af2adc9e22798894021254ef9f5be3a163bfced9" exitCode=0 Dec 03 08:00:01 crc kubenswrapper[4946]: I1203 08:00:01.737604 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" event={"ID":"bbd89e07-ac18-45a6-9713-06cfc08bf40a","Type":"ContainerDied","Data":"779cad687350dfef33e55991af2adc9e22798894021254ef9f5be3a163bfced9"} Dec 03 08:00:01 crc kubenswrapper[4946]: I1203 08:00:01.737820 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" event={"ID":"bbd89e07-ac18-45a6-9713-06cfc08bf40a","Type":"ContainerStarted","Data":"9d772af2a7658bd12e55854725c2ab56e98e7b000db189abbfefb176567b98d6"} Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.090173 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.175867 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd89e07-ac18-45a6-9713-06cfc08bf40a-config-volume\") pod \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.176382 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs7lr\" (UniqueName: \"kubernetes.io/projected/bbd89e07-ac18-45a6-9713-06cfc08bf40a-kube-api-access-xs7lr\") pod \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.176556 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd89e07-ac18-45a6-9713-06cfc08bf40a-secret-volume\") pod \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\" (UID: \"bbd89e07-ac18-45a6-9713-06cfc08bf40a\") " Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.176998 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbd89e07-ac18-45a6-9713-06cfc08bf40a-config-volume" (OuterVolumeSpecName: "config-volume") pod "bbd89e07-ac18-45a6-9713-06cfc08bf40a" (UID: "bbd89e07-ac18-45a6-9713-06cfc08bf40a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.181900 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbd89e07-ac18-45a6-9713-06cfc08bf40a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bbd89e07-ac18-45a6-9713-06cfc08bf40a" (UID: "bbd89e07-ac18-45a6-9713-06cfc08bf40a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.182834 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbd89e07-ac18-45a6-9713-06cfc08bf40a-kube-api-access-xs7lr" (OuterVolumeSpecName: "kube-api-access-xs7lr") pod "bbd89e07-ac18-45a6-9713-06cfc08bf40a" (UID: "bbd89e07-ac18-45a6-9713-06cfc08bf40a"). InnerVolumeSpecName "kube-api-access-xs7lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.278713 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd89e07-ac18-45a6-9713-06cfc08bf40a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.278791 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs7lr\" (UniqueName: \"kubernetes.io/projected/bbd89e07-ac18-45a6-9713-06cfc08bf40a-kube-api-access-xs7lr\") on node \"crc\" DevicePath \"\"" Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.278816 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd89e07-ac18-45a6-9713-06cfc08bf40a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.754690 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" event={"ID":"bbd89e07-ac18-45a6-9713-06cfc08bf40a","Type":"ContainerDied","Data":"9d772af2a7658bd12e55854725c2ab56e98e7b000db189abbfefb176567b98d6"} Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.754787 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d772af2a7658bd12e55854725c2ab56e98e7b000db189abbfefb176567b98d6" Dec 03 08:00:03 crc kubenswrapper[4946]: I1203 08:00:03.754876 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq" Dec 03 08:00:04 crc kubenswrapper[4946]: I1203 08:00:04.167120 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw"] Dec 03 08:00:04 crc kubenswrapper[4946]: I1203 08:00:04.175772 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412435-b9xpw"] Dec 03 08:00:05 crc kubenswrapper[4946]: I1203 08:00:05.609638 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0e9ec0-504f-496e-b9e5-c7aa89352318" path="/var/lib/kubelet/pods/af0e9ec0-504f-496e-b9e5-c7aa89352318/volumes" Dec 03 08:00:13 crc kubenswrapper[4946]: I1203 08:00:13.909064 4946 scope.go:117] "RemoveContainer" containerID="59b2e84655c7126bb52221921ca30a70758edc78fac3c60480dd87b24524d12a" Dec 03 08:01:53 crc kubenswrapper[4946]: I1203 08:01:53.039225 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:01:53 crc kubenswrapper[4946]: I1203 08:01:53.039795 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:02:23 crc kubenswrapper[4946]: I1203 08:02:23.039290 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:02:23 crc kubenswrapper[4946]: I1203 08:02:23.040101 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.039285 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.040204 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.040280 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.041330 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.041464 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" gracePeriod=600 Dec 03 08:02:53 crc kubenswrapper[4946]: E1203 08:02:53.163788 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.434291 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" exitCode=0 Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.434374 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f"} Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.434950 4946 scope.go:117] "RemoveContainer" containerID="ca923913b5c35f3d0179d87c55d99b6dc0fcac16a63f3f56c27146bef1002661" Dec 03 08:02:53 crc kubenswrapper[4946]: I1203 08:02:53.435692 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:02:53 crc kubenswrapper[4946]: E1203 08:02:53.436242 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:03:04 crc kubenswrapper[4946]: I1203 08:03:04.593031 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:03:04 crc kubenswrapper[4946]: E1203 08:03:04.593830 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:03:18 crc kubenswrapper[4946]: I1203 08:03:18.693731 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:03:18 crc kubenswrapper[4946]: E1203 08:03:18.694730 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:03:32 crc kubenswrapper[4946]: I1203 08:03:32.593954 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:03:32 crc kubenswrapper[4946]: E1203 08:03:32.595323 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:03:43 crc kubenswrapper[4946]: I1203 08:03:43.592850 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:03:43 crc kubenswrapper[4946]: E1203 08:03:43.594121 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:03:56 crc kubenswrapper[4946]: I1203 08:03:56.593088 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:03:56 crc kubenswrapper[4946]: E1203 08:03:56.594327 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:04:07 crc kubenswrapper[4946]: I1203 08:04:07.604533 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:04:07 crc kubenswrapper[4946]: E1203 08:04:07.606282 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:04:19 crc kubenswrapper[4946]: I1203 08:04:19.593165 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:04:19 crc kubenswrapper[4946]: E1203 08:04:19.594549 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:04:30 crc kubenswrapper[4946]: I1203 08:04:30.593091 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:04:30 crc kubenswrapper[4946]: E1203 08:04:30.593727 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:04:41 crc kubenswrapper[4946]: I1203 08:04:41.593440 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:04:41 crc kubenswrapper[4946]: E1203 08:04:41.594644 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:04:55 crc kubenswrapper[4946]: I1203 08:04:55.592700 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:04:55 crc kubenswrapper[4946]: E1203 08:04:55.594101 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:05:06 crc kubenswrapper[4946]: I1203 08:05:06.593424 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:05:06 crc kubenswrapper[4946]: E1203 08:05:06.594697 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:05:20 crc kubenswrapper[4946]: I1203 08:05:20.592782 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:05:20 crc kubenswrapper[4946]: E1203 08:05:20.593464 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:05:32 crc kubenswrapper[4946]: I1203 08:05:32.592939 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:05:32 crc kubenswrapper[4946]: E1203 08:05:32.593935 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:05:47 crc kubenswrapper[4946]: I1203 08:05:47.601514 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:05:47 crc kubenswrapper[4946]: E1203 08:05:47.603528 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:06:00 crc kubenswrapper[4946]: I1203 08:06:00.593364 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:06:00 crc kubenswrapper[4946]: E1203 08:06:00.596485 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:06:11 crc kubenswrapper[4946]: I1203 08:06:11.593065 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:06:11 crc kubenswrapper[4946]: E1203 08:06:11.594229 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:06:26 crc kubenswrapper[4946]: I1203 08:06:26.593979 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:06:26 crc kubenswrapper[4946]: E1203 08:06:26.597703 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:06:37 crc kubenswrapper[4946]: I1203 08:06:37.597930 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:06:37 crc kubenswrapper[4946]: E1203 08:06:37.598692 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:06:48 crc kubenswrapper[4946]: I1203 08:06:48.592577 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:06:48 crc kubenswrapper[4946]: E1203 08:06:48.593326 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:06:59 crc kubenswrapper[4946]: I1203 08:06:59.592821 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:06:59 crc kubenswrapper[4946]: E1203 08:06:59.593546 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.657862 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4w5n8"] Dec 03 08:07:07 crc kubenswrapper[4946]: E1203 08:07:07.658992 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd89e07-ac18-45a6-9713-06cfc08bf40a" containerName="collect-profiles" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.659021 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd89e07-ac18-45a6-9713-06cfc08bf40a" containerName="collect-profiles" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.659399 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbd89e07-ac18-45a6-9713-06cfc08bf40a" containerName="collect-profiles" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.664325 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.673657 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4w5n8"] Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.748349 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-utilities\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.748480 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-catalog-content\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.748603 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7f2w\" (UniqueName: \"kubernetes.io/projected/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-kube-api-access-w7f2w\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.850035 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7f2w\" (UniqueName: \"kubernetes.io/projected/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-kube-api-access-w7f2w\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.850132 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-utilities\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.850206 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-catalog-content\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.850812 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-catalog-content\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.850848 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-utilities\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.880528 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7f2w\" (UniqueName: \"kubernetes.io/projected/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-kube-api-access-w7f2w\") pod \"community-operators-4w5n8\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:07 crc kubenswrapper[4946]: I1203 08:07:07.992672 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:08 crc kubenswrapper[4946]: I1203 08:07:08.519167 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4w5n8"] Dec 03 08:07:08 crc kubenswrapper[4946]: W1203 08:07:08.525234 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb86176f_448c_4e96_bec8_3b8e0dcd6b01.slice/crio-669799bd1a3f3c81c3ca78efc36bc836e02c25488a4c75d1b93ee2e1dbed4b38 WatchSource:0}: Error finding container 669799bd1a3f3c81c3ca78efc36bc836e02c25488a4c75d1b93ee2e1dbed4b38: Status 404 returned error can't find the container with id 669799bd1a3f3c81c3ca78efc36bc836e02c25488a4c75d1b93ee2e1dbed4b38 Dec 03 08:07:08 crc kubenswrapper[4946]: I1203 08:07:08.848115 4946 generic.go:334] "Generic (PLEG): container finished" podID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerID="c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6" exitCode=0 Dec 03 08:07:08 crc kubenswrapper[4946]: I1203 08:07:08.848213 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4w5n8" event={"ID":"cb86176f-448c-4e96-bec8-3b8e0dcd6b01","Type":"ContainerDied","Data":"c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6"} Dec 03 08:07:08 crc kubenswrapper[4946]: I1203 08:07:08.848566 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4w5n8" event={"ID":"cb86176f-448c-4e96-bec8-3b8e0dcd6b01","Type":"ContainerStarted","Data":"669799bd1a3f3c81c3ca78efc36bc836e02c25488a4c75d1b93ee2e1dbed4b38"} Dec 03 08:07:08 crc kubenswrapper[4946]: I1203 08:07:08.850441 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.032054 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dgrrc"] Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.035277 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.044602 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgrrc"] Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.082876 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-catalog-content\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.083013 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4t8c\" (UniqueName: \"kubernetes.io/projected/be101e5f-5946-4c42-8304-de6110d78ce2-kube-api-access-v4t8c\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.083077 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-utilities\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.184850 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-catalog-content\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.184989 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4t8c\" (UniqueName: \"kubernetes.io/projected/be101e5f-5946-4c42-8304-de6110d78ce2-kube-api-access-v4t8c\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.185088 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-utilities\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.185337 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-catalog-content\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.185635 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-utilities\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.210809 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4t8c\" (UniqueName: \"kubernetes.io/projected/be101e5f-5946-4c42-8304-de6110d78ce2-kube-api-access-v4t8c\") pod \"redhat-marketplace-dgrrc\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.355456 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.582557 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgrrc"] Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.863295 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgrrc" event={"ID":"be101e5f-5946-4c42-8304-de6110d78ce2","Type":"ContainerStarted","Data":"2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe"} Dec 03 08:07:09 crc kubenswrapper[4946]: I1203 08:07:09.863682 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgrrc" event={"ID":"be101e5f-5946-4c42-8304-de6110d78ce2","Type":"ContainerStarted","Data":"8d29df788ea46a0531b225ec5bcbd24045ec690c5fa9988b7504dd857ee65768"} Dec 03 08:07:10 crc kubenswrapper[4946]: I1203 08:07:10.593287 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:07:10 crc kubenswrapper[4946]: E1203 08:07:10.593564 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:07:10 crc kubenswrapper[4946]: I1203 08:07:10.873091 4946 generic.go:334] "Generic (PLEG): container finished" podID="be101e5f-5946-4c42-8304-de6110d78ce2" containerID="2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe" exitCode=0 Dec 03 08:07:10 crc kubenswrapper[4946]: I1203 08:07:10.873154 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgrrc" event={"ID":"be101e5f-5946-4c42-8304-de6110d78ce2","Type":"ContainerDied","Data":"2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe"} Dec 03 08:07:10 crc kubenswrapper[4946]: I1203 08:07:10.875694 4946 generic.go:334] "Generic (PLEG): container finished" podID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerID="1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525" exitCode=0 Dec 03 08:07:10 crc kubenswrapper[4946]: I1203 08:07:10.875792 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4w5n8" event={"ID":"cb86176f-448c-4e96-bec8-3b8e0dcd6b01","Type":"ContainerDied","Data":"1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525"} Dec 03 08:07:11 crc kubenswrapper[4946]: I1203 08:07:11.885944 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4w5n8" event={"ID":"cb86176f-448c-4e96-bec8-3b8e0dcd6b01","Type":"ContainerStarted","Data":"3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09"} Dec 03 08:07:11 crc kubenswrapper[4946]: I1203 08:07:11.911221 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4w5n8" podStartSLOduration=2.43456101 podStartE2EDuration="4.911198192s" podCreationTimestamp="2025-12-03 08:07:07 +0000 UTC" firstStartedPulling="2025-12-03 08:07:08.850150477 +0000 UTC m=+4621.646840596" lastFinishedPulling="2025-12-03 08:07:11.326787609 +0000 UTC m=+4624.123477778" observedRunningTime="2025-12-03 08:07:11.905677586 +0000 UTC m=+4624.702367715" watchObservedRunningTime="2025-12-03 08:07:11.911198192 +0000 UTC m=+4624.707888341" Dec 03 08:07:12 crc kubenswrapper[4946]: I1203 08:07:12.896806 4946 generic.go:334] "Generic (PLEG): container finished" podID="be101e5f-5946-4c42-8304-de6110d78ce2" containerID="594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9" exitCode=0 Dec 03 08:07:12 crc kubenswrapper[4946]: I1203 08:07:12.896915 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgrrc" event={"ID":"be101e5f-5946-4c42-8304-de6110d78ce2","Type":"ContainerDied","Data":"594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9"} Dec 03 08:07:13 crc kubenswrapper[4946]: I1203 08:07:13.907269 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgrrc" event={"ID":"be101e5f-5946-4c42-8304-de6110d78ce2","Type":"ContainerStarted","Data":"a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5"} Dec 03 08:07:13 crc kubenswrapper[4946]: I1203 08:07:13.929226 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dgrrc" podStartSLOduration=2.503793275 podStartE2EDuration="5.929190027s" podCreationTimestamp="2025-12-03 08:07:08 +0000 UTC" firstStartedPulling="2025-12-03 08:07:09.865282596 +0000 UTC m=+4622.661972755" lastFinishedPulling="2025-12-03 08:07:13.290679378 +0000 UTC m=+4626.087369507" observedRunningTime="2025-12-03 08:07:13.928390005 +0000 UTC m=+4626.725080154" watchObservedRunningTime="2025-12-03 08:07:13.929190027 +0000 UTC m=+4626.725880206" Dec 03 08:07:18 crc kubenswrapper[4946]: I1203 08:07:18.061714 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:18 crc kubenswrapper[4946]: I1203 08:07:18.062386 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:18 crc kubenswrapper[4946]: I1203 08:07:18.126228 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:19 crc kubenswrapper[4946]: I1203 08:07:19.157027 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:19 crc kubenswrapper[4946]: I1203 08:07:19.228528 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4w5n8"] Dec 03 08:07:19 crc kubenswrapper[4946]: I1203 08:07:19.356933 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:19 crc kubenswrapper[4946]: I1203 08:07:19.357053 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:19 crc kubenswrapper[4946]: I1203 08:07:19.435082 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:20 crc kubenswrapper[4946]: I1203 08:07:20.167635 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:20 crc kubenswrapper[4946]: I1203 08:07:20.811971 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgrrc"] Dec 03 08:07:21 crc kubenswrapper[4946]: I1203 08:07:21.123023 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4w5n8" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerName="registry-server" containerID="cri-o://3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09" gracePeriod=2 Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.139114 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dgrrc" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" containerName="registry-server" containerID="cri-o://a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5" gracePeriod=2 Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.590715 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.653583 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.749229 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-utilities\") pod \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.749395 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-catalog-content\") pod \"be101e5f-5946-4c42-8304-de6110d78ce2\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.749458 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4t8c\" (UniqueName: \"kubernetes.io/projected/be101e5f-5946-4c42-8304-de6110d78ce2-kube-api-access-v4t8c\") pod \"be101e5f-5946-4c42-8304-de6110d78ce2\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.749509 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7f2w\" (UniqueName: \"kubernetes.io/projected/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-kube-api-access-w7f2w\") pod \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.749619 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-catalog-content\") pod \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\" (UID: \"cb86176f-448c-4e96-bec8-3b8e0dcd6b01\") " Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.749655 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-utilities\") pod \"be101e5f-5946-4c42-8304-de6110d78ce2\" (UID: \"be101e5f-5946-4c42-8304-de6110d78ce2\") " Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.750114 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-utilities" (OuterVolumeSpecName: "utilities") pod "cb86176f-448c-4e96-bec8-3b8e0dcd6b01" (UID: "cb86176f-448c-4e96-bec8-3b8e0dcd6b01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.752551 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-utilities" (OuterVolumeSpecName: "utilities") pod "be101e5f-5946-4c42-8304-de6110d78ce2" (UID: "be101e5f-5946-4c42-8304-de6110d78ce2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.755728 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-kube-api-access-w7f2w" (OuterVolumeSpecName: "kube-api-access-w7f2w") pod "cb86176f-448c-4e96-bec8-3b8e0dcd6b01" (UID: "cb86176f-448c-4e96-bec8-3b8e0dcd6b01"). InnerVolumeSpecName "kube-api-access-w7f2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.756192 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be101e5f-5946-4c42-8304-de6110d78ce2-kube-api-access-v4t8c" (OuterVolumeSpecName: "kube-api-access-v4t8c") pod "be101e5f-5946-4c42-8304-de6110d78ce2" (UID: "be101e5f-5946-4c42-8304-de6110d78ce2"). InnerVolumeSpecName "kube-api-access-v4t8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.767694 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be101e5f-5946-4c42-8304-de6110d78ce2" (UID: "be101e5f-5946-4c42-8304-de6110d78ce2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.807758 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb86176f-448c-4e96-bec8-3b8e0dcd6b01" (UID: "cb86176f-448c-4e96-bec8-3b8e0dcd6b01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.851769 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.851866 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.851890 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4t8c\" (UniqueName: \"kubernetes.io/projected/be101e5f-5946-4c42-8304-de6110d78ce2-kube-api-access-v4t8c\") on node \"crc\" DevicePath \"\"" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.851908 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7f2w\" (UniqueName: \"kubernetes.io/projected/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-kube-api-access-w7f2w\") on node \"crc\" DevicePath \"\"" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.851930 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb86176f-448c-4e96-bec8-3b8e0dcd6b01-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:07:22 crc kubenswrapper[4946]: I1203 08:07:22.851946 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be101e5f-5946-4c42-8304-de6110d78ce2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.148977 4946 generic.go:334] "Generic (PLEG): container finished" podID="be101e5f-5946-4c42-8304-de6110d78ce2" containerID="a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5" exitCode=0 Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.149025 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgrrc" event={"ID":"be101e5f-5946-4c42-8304-de6110d78ce2","Type":"ContainerDied","Data":"a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5"} Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.149061 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgrrc" event={"ID":"be101e5f-5946-4c42-8304-de6110d78ce2","Type":"ContainerDied","Data":"8d29df788ea46a0531b225ec5bcbd24045ec690c5fa9988b7504dd857ee65768"} Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.149080 4946 scope.go:117] "RemoveContainer" containerID="a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.149239 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgrrc" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.155069 4946 generic.go:334] "Generic (PLEG): container finished" podID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerID="3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09" exitCode=0 Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.155108 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4w5n8" event={"ID":"cb86176f-448c-4e96-bec8-3b8e0dcd6b01","Type":"ContainerDied","Data":"3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09"} Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.155132 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4w5n8" event={"ID":"cb86176f-448c-4e96-bec8-3b8e0dcd6b01","Type":"ContainerDied","Data":"669799bd1a3f3c81c3ca78efc36bc836e02c25488a4c75d1b93ee2e1dbed4b38"} Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.155172 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4w5n8" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.171475 4946 scope.go:117] "RemoveContainer" containerID="594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.196521 4946 scope.go:117] "RemoveContainer" containerID="2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.201504 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgrrc"] Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.206333 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgrrc"] Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.226665 4946 scope.go:117] "RemoveContainer" containerID="a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5" Dec 03 08:07:23 crc kubenswrapper[4946]: E1203 08:07:23.227416 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5\": container with ID starting with a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5 not found: ID does not exist" containerID="a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.227474 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5"} err="failed to get container status \"a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5\": rpc error: code = NotFound desc = could not find container \"a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5\": container with ID starting with a2614062ee46bd519fb370691638a866e7f8ad261bb6ababcdb0ad5c18a70cd5 not found: ID does not exist" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.227521 4946 scope.go:117] "RemoveContainer" containerID="594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9" Dec 03 08:07:23 crc kubenswrapper[4946]: E1203 08:07:23.228063 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9\": container with ID starting with 594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9 not found: ID does not exist" containerID="594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.228125 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9"} err="failed to get container status \"594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9\": rpc error: code = NotFound desc = could not find container \"594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9\": container with ID starting with 594be07a36657d50f910f389c0a8be579d7cc0d0681357be88f3b906016a94a9 not found: ID does not exist" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.228176 4946 scope.go:117] "RemoveContainer" containerID="2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe" Dec 03 08:07:23 crc kubenswrapper[4946]: E1203 08:07:23.228824 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe\": container with ID starting with 2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe not found: ID does not exist" containerID="2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.228854 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe"} err="failed to get container status \"2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe\": rpc error: code = NotFound desc = could not find container \"2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe\": container with ID starting with 2d1c2e86b0d3c45cc79175203776d486c366370e09d9fba920e3a813d78a1bfe not found: ID does not exist" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.228876 4946 scope.go:117] "RemoveContainer" containerID="3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.232063 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4w5n8"] Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.239757 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4w5n8"] Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.252459 4946 scope.go:117] "RemoveContainer" containerID="1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.273511 4946 scope.go:117] "RemoveContainer" containerID="c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.296409 4946 scope.go:117] "RemoveContainer" containerID="3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09" Dec 03 08:07:23 crc kubenswrapper[4946]: E1203 08:07:23.297116 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09\": container with ID starting with 3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09 not found: ID does not exist" containerID="3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.297156 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09"} err="failed to get container status \"3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09\": rpc error: code = NotFound desc = could not find container \"3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09\": container with ID starting with 3a908bc803236781f5a7f7dbf0a3be8e23f309919255c32fb8610230a3066c09 not found: ID does not exist" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.297180 4946 scope.go:117] "RemoveContainer" containerID="1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525" Dec 03 08:07:23 crc kubenswrapper[4946]: E1203 08:07:23.297528 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525\": container with ID starting with 1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525 not found: ID does not exist" containerID="1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.297555 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525"} err="failed to get container status \"1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525\": rpc error: code = NotFound desc = could not find container \"1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525\": container with ID starting with 1e7c01aaa94a5075cac514b6b7b8725ec95ae6f7c2984c0b67c24b15ae561525 not found: ID does not exist" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.297578 4946 scope.go:117] "RemoveContainer" containerID="c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6" Dec 03 08:07:23 crc kubenswrapper[4946]: E1203 08:07:23.298005 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6\": container with ID starting with c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6 not found: ID does not exist" containerID="c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.298033 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6"} err="failed to get container status \"c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6\": rpc error: code = NotFound desc = could not find container \"c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6\": container with ID starting with c9e0bc1b92171bb5ec66d2fb2c2bce0a72b8e6af66560d937ba71b0e9505aaa6 not found: ID does not exist" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.602699 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" path="/var/lib/kubelet/pods/be101e5f-5946-4c42-8304-de6110d78ce2/volumes" Dec 03 08:07:23 crc kubenswrapper[4946]: I1203 08:07:23.603616 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" path="/var/lib/kubelet/pods/cb86176f-448c-4e96-bec8-3b8e0dcd6b01/volumes" Dec 03 08:07:25 crc kubenswrapper[4946]: I1203 08:07:25.594068 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:07:25 crc kubenswrapper[4946]: E1203 08:07:25.594896 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:07:37 crc kubenswrapper[4946]: I1203 08:07:37.592988 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:07:37 crc kubenswrapper[4946]: E1203 08:07:37.593928 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:07:52 crc kubenswrapper[4946]: I1203 08:07:52.593239 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:07:52 crc kubenswrapper[4946]: E1203 08:07:52.594832 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:08:07 crc kubenswrapper[4946]: I1203 08:08:07.600790 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:08:08 crc kubenswrapper[4946]: I1203 08:08:08.585225 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"6ee6f09c62da9ecbff81cbd128729bb913930c8b592dcef4007818ed9ee04120"} Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.128634 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gfd2n"] Dec 03 08:09:11 crc kubenswrapper[4946]: E1203 08:09:11.129914 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerName="registry-server" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.130281 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerName="registry-server" Dec 03 08:09:11 crc kubenswrapper[4946]: E1203 08:09:11.130309 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" containerName="extract-utilities" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.130319 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" containerName="extract-utilities" Dec 03 08:09:11 crc kubenswrapper[4946]: E1203 08:09:11.130339 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerName="extract-utilities" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.130352 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerName="extract-utilities" Dec 03 08:09:11 crc kubenswrapper[4946]: E1203 08:09:11.130370 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" containerName="registry-server" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.130380 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" containerName="registry-server" Dec 03 08:09:11 crc kubenswrapper[4946]: E1203 08:09:11.130434 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerName="extract-content" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.130442 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerName="extract-content" Dec 03 08:09:11 crc kubenswrapper[4946]: E1203 08:09:11.130476 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" containerName="extract-content" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.130485 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" containerName="extract-content" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.130890 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="be101e5f-5946-4c42-8304-de6110d78ce2" containerName="registry-server" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.130911 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb86176f-448c-4e96-bec8-3b8e0dcd6b01" containerName="registry-server" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.136435 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.144198 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gfd2n"] Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.237082 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-catalog-content\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.237355 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8pzl\" (UniqueName: \"kubernetes.io/projected/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-kube-api-access-k8pzl\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.237672 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-utilities\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.338910 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-utilities\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.338988 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-catalog-content\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.339013 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8pzl\" (UniqueName: \"kubernetes.io/projected/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-kube-api-access-k8pzl\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.339573 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-utilities\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.339976 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-catalog-content\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.361783 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8pzl\" (UniqueName: \"kubernetes.io/projected/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-kube-api-access-k8pzl\") pod \"certified-operators-gfd2n\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.502985 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:11 crc kubenswrapper[4946]: I1203 08:09:11.794476 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gfd2n"] Dec 03 08:09:12 crc kubenswrapper[4946]: I1203 08:09:12.200561 4946 generic.go:334] "Generic (PLEG): container finished" podID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerID="14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54" exitCode=0 Dec 03 08:09:12 crc kubenswrapper[4946]: I1203 08:09:12.200628 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfd2n" event={"ID":"3be63d6f-a3f0-4342-bd5a-c22b3e044e87","Type":"ContainerDied","Data":"14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54"} Dec 03 08:09:12 crc kubenswrapper[4946]: I1203 08:09:12.200686 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfd2n" event={"ID":"3be63d6f-a3f0-4342-bd5a-c22b3e044e87","Type":"ContainerStarted","Data":"c90ab30471bf598e92f388515bde6d9604a4adfd71e3d2da02a88bdb8f064a98"} Dec 03 08:09:13 crc kubenswrapper[4946]: I1203 08:09:13.209921 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfd2n" event={"ID":"3be63d6f-a3f0-4342-bd5a-c22b3e044e87","Type":"ContainerStarted","Data":"ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031"} Dec 03 08:09:14 crc kubenswrapper[4946]: I1203 08:09:14.219294 4946 generic.go:334] "Generic (PLEG): container finished" podID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerID="ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031" exitCode=0 Dec 03 08:09:14 crc kubenswrapper[4946]: I1203 08:09:14.219380 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfd2n" event={"ID":"3be63d6f-a3f0-4342-bd5a-c22b3e044e87","Type":"ContainerDied","Data":"ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031"} Dec 03 08:09:15 crc kubenswrapper[4946]: I1203 08:09:15.231664 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfd2n" event={"ID":"3be63d6f-a3f0-4342-bd5a-c22b3e044e87","Type":"ContainerStarted","Data":"006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe"} Dec 03 08:09:15 crc kubenswrapper[4946]: I1203 08:09:15.269091 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gfd2n" podStartSLOduration=1.807368013 podStartE2EDuration="4.269034148s" podCreationTimestamp="2025-12-03 08:09:11 +0000 UTC" firstStartedPulling="2025-12-03 08:09:12.204328406 +0000 UTC m=+4745.001018545" lastFinishedPulling="2025-12-03 08:09:14.665994531 +0000 UTC m=+4747.462684680" observedRunningTime="2025-12-03 08:09:15.259409273 +0000 UTC m=+4748.056099472" watchObservedRunningTime="2025-12-03 08:09:15.269034148 +0000 UTC m=+4748.065724297" Dec 03 08:09:21 crc kubenswrapper[4946]: I1203 08:09:21.503547 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:21 crc kubenswrapper[4946]: I1203 08:09:21.504097 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:21 crc kubenswrapper[4946]: I1203 08:09:21.566879 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:22 crc kubenswrapper[4946]: I1203 08:09:22.352110 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:22 crc kubenswrapper[4946]: I1203 08:09:22.425186 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gfd2n"] Dec 03 08:09:24 crc kubenswrapper[4946]: I1203 08:09:24.318576 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gfd2n" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerName="registry-server" containerID="cri-o://006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe" gracePeriod=2 Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.308416 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.337613 4946 generic.go:334] "Generic (PLEG): container finished" podID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerID="006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe" exitCode=0 Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.337667 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfd2n" event={"ID":"3be63d6f-a3f0-4342-bd5a-c22b3e044e87","Type":"ContainerDied","Data":"006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe"} Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.337702 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfd2n" event={"ID":"3be63d6f-a3f0-4342-bd5a-c22b3e044e87","Type":"ContainerDied","Data":"c90ab30471bf598e92f388515bde6d9604a4adfd71e3d2da02a88bdb8f064a98"} Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.337724 4946 scope.go:117] "RemoveContainer" containerID="006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.337999 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gfd2n" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.369280 4946 scope.go:117] "RemoveContainer" containerID="ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.396541 4946 scope.go:117] "RemoveContainer" containerID="14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.414587 4946 scope.go:117] "RemoveContainer" containerID="006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe" Dec 03 08:09:25 crc kubenswrapper[4946]: E1203 08:09:25.415195 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe\": container with ID starting with 006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe not found: ID does not exist" containerID="006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.415250 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe"} err="failed to get container status \"006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe\": rpc error: code = NotFound desc = could not find container \"006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe\": container with ID starting with 006647c264441975fc4c1daa660712336dd3b126216068ea9454b1f8b07df1fe not found: ID does not exist" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.415431 4946 scope.go:117] "RemoveContainer" containerID="ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031" Dec 03 08:09:25 crc kubenswrapper[4946]: E1203 08:09:25.415796 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031\": container with ID starting with ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031 not found: ID does not exist" containerID="ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.415838 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031"} err="failed to get container status \"ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031\": rpc error: code = NotFound desc = could not find container \"ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031\": container with ID starting with ec76fa85762ee3fbaadddc590ee019a2662ccb5ec6ba896ef959d012c944a031 not found: ID does not exist" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.415866 4946 scope.go:117] "RemoveContainer" containerID="14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54" Dec 03 08:09:25 crc kubenswrapper[4946]: E1203 08:09:25.416187 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54\": container with ID starting with 14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54 not found: ID does not exist" containerID="14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.416261 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54"} err="failed to get container status \"14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54\": rpc error: code = NotFound desc = could not find container \"14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54\": container with ID starting with 14b636cda5154d435643aa46cea30175118f9ea81e792ffb4bf6820b5fcfad54 not found: ID does not exist" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.476813 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8pzl\" (UniqueName: \"kubernetes.io/projected/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-kube-api-access-k8pzl\") pod \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.476994 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-utilities\") pod \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.477128 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-catalog-content\") pod \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\" (UID: \"3be63d6f-a3f0-4342-bd5a-c22b3e044e87\") " Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.478275 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-utilities" (OuterVolumeSpecName: "utilities") pod "3be63d6f-a3f0-4342-bd5a-c22b3e044e87" (UID: "3be63d6f-a3f0-4342-bd5a-c22b3e044e87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.482585 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-kube-api-access-k8pzl" (OuterVolumeSpecName: "kube-api-access-k8pzl") pod "3be63d6f-a3f0-4342-bd5a-c22b3e044e87" (UID: "3be63d6f-a3f0-4342-bd5a-c22b3e044e87"). InnerVolumeSpecName "kube-api-access-k8pzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.551949 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3be63d6f-a3f0-4342-bd5a-c22b3e044e87" (UID: "3be63d6f-a3f0-4342-bd5a-c22b3e044e87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.579782 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8pzl\" (UniqueName: \"kubernetes.io/projected/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-kube-api-access-k8pzl\") on node \"crc\" DevicePath \"\"" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.579833 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.579854 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3be63d6f-a3f0-4342-bd5a-c22b3e044e87-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.675849 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gfd2n"] Dec 03 08:09:25 crc kubenswrapper[4946]: I1203 08:09:25.688246 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gfd2n"] Dec 03 08:09:27 crc kubenswrapper[4946]: I1203 08:09:27.610160 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" path="/var/lib/kubelet/pods/3be63d6f-a3f0-4342-bd5a-c22b3e044e87/volumes" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.404321 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fjpcd"] Dec 03 08:10:04 crc kubenswrapper[4946]: E1203 08:10:04.405699 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerName="extract-utilities" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.405732 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerName="extract-utilities" Dec 03 08:10:04 crc kubenswrapper[4946]: E1203 08:10:04.405811 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerName="registry-server" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.405831 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerName="registry-server" Dec 03 08:10:04 crc kubenswrapper[4946]: E1203 08:10:04.405856 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerName="extract-content" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.405870 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerName="extract-content" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.406171 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="3be63d6f-a3f0-4342-bd5a-c22b3e044e87" containerName="registry-server" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.408049 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.423440 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjpcd"] Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.478986 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4pjk\" (UniqueName: \"kubernetes.io/projected/0308b307-ec82-49fd-b325-1c40da207328-kube-api-access-m4pjk\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.479266 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-utilities\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.479393 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-catalog-content\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.581470 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4pjk\" (UniqueName: \"kubernetes.io/projected/0308b307-ec82-49fd-b325-1c40da207328-kube-api-access-m4pjk\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.581872 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-utilities\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.582079 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-catalog-content\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.582400 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-utilities\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.582439 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-catalog-content\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.603289 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4pjk\" (UniqueName: \"kubernetes.io/projected/0308b307-ec82-49fd-b325-1c40da207328-kube-api-access-m4pjk\") pod \"redhat-operators-fjpcd\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:04 crc kubenswrapper[4946]: I1203 08:10:04.757361 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:05 crc kubenswrapper[4946]: I1203 08:10:05.231274 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjpcd"] Dec 03 08:10:05 crc kubenswrapper[4946]: I1203 08:10:05.724923 4946 generic.go:334] "Generic (PLEG): container finished" podID="0308b307-ec82-49fd-b325-1c40da207328" containerID="4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8" exitCode=0 Dec 03 08:10:05 crc kubenswrapper[4946]: I1203 08:10:05.724971 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjpcd" event={"ID":"0308b307-ec82-49fd-b325-1c40da207328","Type":"ContainerDied","Data":"4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8"} Dec 03 08:10:05 crc kubenswrapper[4946]: I1203 08:10:05.725026 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjpcd" event={"ID":"0308b307-ec82-49fd-b325-1c40da207328","Type":"ContainerStarted","Data":"19b936fc294569b0d6f1b2615fac93080ac9cf24c78f87c00d871fdd18d4b8bd"} Dec 03 08:10:07 crc kubenswrapper[4946]: I1203 08:10:07.745715 4946 generic.go:334] "Generic (PLEG): container finished" podID="0308b307-ec82-49fd-b325-1c40da207328" containerID="9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4" exitCode=0 Dec 03 08:10:07 crc kubenswrapper[4946]: I1203 08:10:07.746233 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjpcd" event={"ID":"0308b307-ec82-49fd-b325-1c40da207328","Type":"ContainerDied","Data":"9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4"} Dec 03 08:10:08 crc kubenswrapper[4946]: I1203 08:10:08.767061 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjpcd" event={"ID":"0308b307-ec82-49fd-b325-1c40da207328","Type":"ContainerStarted","Data":"2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e"} Dec 03 08:10:08 crc kubenswrapper[4946]: I1203 08:10:08.791379 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fjpcd" podStartSLOduration=2.354511777 podStartE2EDuration="4.791326943s" podCreationTimestamp="2025-12-03 08:10:04 +0000 UTC" firstStartedPulling="2025-12-03 08:10:05.726471516 +0000 UTC m=+4798.523161625" lastFinishedPulling="2025-12-03 08:10:08.163286682 +0000 UTC m=+4800.959976791" observedRunningTime="2025-12-03 08:10:08.788614541 +0000 UTC m=+4801.585304650" watchObservedRunningTime="2025-12-03 08:10:08.791326943 +0000 UTC m=+4801.588017062" Dec 03 08:10:14 crc kubenswrapper[4946]: I1203 08:10:14.758518 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:14 crc kubenswrapper[4946]: I1203 08:10:14.759272 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:14 crc kubenswrapper[4946]: I1203 08:10:14.833331 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:14 crc kubenswrapper[4946]: I1203 08:10:14.896889 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:15 crc kubenswrapper[4946]: I1203 08:10:15.102140 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjpcd"] Dec 03 08:10:16 crc kubenswrapper[4946]: I1203 08:10:16.831439 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fjpcd" podUID="0308b307-ec82-49fd-b325-1c40da207328" containerName="registry-server" containerID="cri-o://2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e" gracePeriod=2 Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.248424 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.387916 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4pjk\" (UniqueName: \"kubernetes.io/projected/0308b307-ec82-49fd-b325-1c40da207328-kube-api-access-m4pjk\") pod \"0308b307-ec82-49fd-b325-1c40da207328\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.387973 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-catalog-content\") pod \"0308b307-ec82-49fd-b325-1c40da207328\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.388071 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-utilities\") pod \"0308b307-ec82-49fd-b325-1c40da207328\" (UID: \"0308b307-ec82-49fd-b325-1c40da207328\") " Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.389036 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-utilities" (OuterVolumeSpecName: "utilities") pod "0308b307-ec82-49fd-b325-1c40da207328" (UID: "0308b307-ec82-49fd-b325-1c40da207328"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.470362 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0308b307-ec82-49fd-b325-1c40da207328-kube-api-access-m4pjk" (OuterVolumeSpecName: "kube-api-access-m4pjk") pod "0308b307-ec82-49fd-b325-1c40da207328" (UID: "0308b307-ec82-49fd-b325-1c40da207328"). InnerVolumeSpecName "kube-api-access-m4pjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.489399 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4pjk\" (UniqueName: \"kubernetes.io/projected/0308b307-ec82-49fd-b325-1c40da207328-kube-api-access-m4pjk\") on node \"crc\" DevicePath \"\"" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.489429 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.536457 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0308b307-ec82-49fd-b325-1c40da207328" (UID: "0308b307-ec82-49fd-b325-1c40da207328"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.590962 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0308b307-ec82-49fd-b325-1c40da207328-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.843314 4946 generic.go:334] "Generic (PLEG): container finished" podID="0308b307-ec82-49fd-b325-1c40da207328" containerID="2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e" exitCode=0 Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.843393 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjpcd" event={"ID":"0308b307-ec82-49fd-b325-1c40da207328","Type":"ContainerDied","Data":"2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e"} Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.843423 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjpcd" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.843818 4946 scope.go:117] "RemoveContainer" containerID="2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.843789 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjpcd" event={"ID":"0308b307-ec82-49fd-b325-1c40da207328","Type":"ContainerDied","Data":"19b936fc294569b0d6f1b2615fac93080ac9cf24c78f87c00d871fdd18d4b8bd"} Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.875075 4946 scope.go:117] "RemoveContainer" containerID="9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.885228 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjpcd"] Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.898535 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fjpcd"] Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.900258 4946 scope.go:117] "RemoveContainer" containerID="4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.925003 4946 scope.go:117] "RemoveContainer" containerID="2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e" Dec 03 08:10:17 crc kubenswrapper[4946]: E1203 08:10:17.925965 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e\": container with ID starting with 2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e not found: ID does not exist" containerID="2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.926130 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e"} err="failed to get container status \"2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e\": rpc error: code = NotFound desc = could not find container \"2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e\": container with ID starting with 2a623fca6f24e1ccab2748b20b1f2c639235aaece6a2cbd9080c43b04a0b426e not found: ID does not exist" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.926241 4946 scope.go:117] "RemoveContainer" containerID="9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4" Dec 03 08:10:17 crc kubenswrapper[4946]: E1203 08:10:17.926878 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4\": container with ID starting with 9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4 not found: ID does not exist" containerID="9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.926951 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4"} err="failed to get container status \"9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4\": rpc error: code = NotFound desc = could not find container \"9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4\": container with ID starting with 9cb034b2ad6b7d580394af0274917e94755e6c106bdbc760a1e8a1d1129946c4 not found: ID does not exist" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.926984 4946 scope.go:117] "RemoveContainer" containerID="4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8" Dec 03 08:10:17 crc kubenswrapper[4946]: E1203 08:10:17.927561 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8\": container with ID starting with 4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8 not found: ID does not exist" containerID="4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8" Dec 03 08:10:17 crc kubenswrapper[4946]: I1203 08:10:17.927655 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8"} err="failed to get container status \"4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8\": rpc error: code = NotFound desc = could not find container \"4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8\": container with ID starting with 4098990ce7ff1bd36ad2939b127f9700399e3ef02f077538408e7d4aa46f27d8 not found: ID does not exist" Dec 03 08:10:19 crc kubenswrapper[4946]: I1203 08:10:19.609246 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0308b307-ec82-49fd-b325-1c40da207328" path="/var/lib/kubelet/pods/0308b307-ec82-49fd-b325-1c40da207328/volumes" Dec 03 08:10:23 crc kubenswrapper[4946]: I1203 08:10:23.040208 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:10:23 crc kubenswrapper[4946]: I1203 08:10:23.040938 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:10:53 crc kubenswrapper[4946]: I1203 08:10:53.039666 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:10:53 crc kubenswrapper[4946]: I1203 08:10:53.041962 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.039079 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.039789 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.039851 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.040671 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ee6f09c62da9ecbff81cbd128729bb913930c8b592dcef4007818ed9ee04120"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.040806 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://6ee6f09c62da9ecbff81cbd128729bb913930c8b592dcef4007818ed9ee04120" gracePeriod=600 Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.491556 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="6ee6f09c62da9ecbff81cbd128729bb913930c8b592dcef4007818ed9ee04120" exitCode=0 Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.491647 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"6ee6f09c62da9ecbff81cbd128729bb913930c8b592dcef4007818ed9ee04120"} Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.492013 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b"} Dec 03 08:11:23 crc kubenswrapper[4946]: I1203 08:11:23.492042 4946 scope.go:117] "RemoveContainer" containerID="63b5b839dd7d780ee4ec1190e6a224012b500efabb83bb70c920efc97c8db64f" Dec 03 08:13:23 crc kubenswrapper[4946]: I1203 08:13:23.040151 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:13:23 crc kubenswrapper[4946]: I1203 08:13:23.040839 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:13:53 crc kubenswrapper[4946]: I1203 08:13:53.039585 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:13:53 crc kubenswrapper[4946]: I1203 08:13:53.040275 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.039220 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.040020 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.040133 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.041140 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.041272 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" gracePeriod=600 Dec 03 08:14:23 crc kubenswrapper[4946]: E1203 08:14:23.174069 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.208401 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" exitCode=0 Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.208496 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b"} Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.208582 4946 scope.go:117] "RemoveContainer" containerID="6ee6f09c62da9ecbff81cbd128729bb913930c8b592dcef4007818ed9ee04120" Dec 03 08:14:23 crc kubenswrapper[4946]: I1203 08:14:23.209258 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:14:23 crc kubenswrapper[4946]: E1203 08:14:23.209529 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:14:36 crc kubenswrapper[4946]: I1203 08:14:36.593234 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:14:36 crc kubenswrapper[4946]: E1203 08:14:36.593953 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:14:51 crc kubenswrapper[4946]: I1203 08:14:51.593892 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:14:51 crc kubenswrapper[4946]: E1203 08:14:51.594939 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.164328 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4"] Dec 03 08:15:00 crc kubenswrapper[4946]: E1203 08:15:00.165446 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0308b307-ec82-49fd-b325-1c40da207328" containerName="registry-server" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.165468 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0308b307-ec82-49fd-b325-1c40da207328" containerName="registry-server" Dec 03 08:15:00 crc kubenswrapper[4946]: E1203 08:15:00.165483 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0308b307-ec82-49fd-b325-1c40da207328" containerName="extract-utilities" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.165494 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0308b307-ec82-49fd-b325-1c40da207328" containerName="extract-utilities" Dec 03 08:15:00 crc kubenswrapper[4946]: E1203 08:15:00.165515 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0308b307-ec82-49fd-b325-1c40da207328" containerName="extract-content" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.165526 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0308b307-ec82-49fd-b325-1c40da207328" containerName="extract-content" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.165784 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0308b307-ec82-49fd-b325-1c40da207328" containerName="registry-server" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.166538 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.170523 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.170897 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.177144 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4"] Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.299927 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-config-volume\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.299996 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr4lr\" (UniqueName: \"kubernetes.io/projected/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-kube-api-access-dr4lr\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.300021 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-secret-volume\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.401523 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-config-volume\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.401846 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr4lr\" (UniqueName: \"kubernetes.io/projected/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-kube-api-access-dr4lr\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.401942 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-secret-volume\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.403243 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-config-volume\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.414448 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-secret-volume\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.421426 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr4lr\" (UniqueName: \"kubernetes.io/projected/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-kube-api-access-dr4lr\") pod \"collect-profiles-29412495-xnwg4\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.494650 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:00 crc kubenswrapper[4946]: I1203 08:15:00.941036 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4"] Dec 03 08:15:01 crc kubenswrapper[4946]: I1203 08:15:01.553283 4946 generic.go:334] "Generic (PLEG): container finished" podID="8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe" containerID="6a75dd080ac098440b8fad4543dfebd6d569a0265d2431ebcf2b13b3f8ec4a3b" exitCode=0 Dec 03 08:15:01 crc kubenswrapper[4946]: I1203 08:15:01.553337 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" event={"ID":"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe","Type":"ContainerDied","Data":"6a75dd080ac098440b8fad4543dfebd6d569a0265d2431ebcf2b13b3f8ec4a3b"} Dec 03 08:15:01 crc kubenswrapper[4946]: I1203 08:15:01.553370 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" event={"ID":"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe","Type":"ContainerStarted","Data":"0d565f066ee5943b8b53e69926bacd8559c64850b35827bf1264f472aff33109"} Dec 03 08:15:02 crc kubenswrapper[4946]: I1203 08:15:02.879858 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.067304 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr4lr\" (UniqueName: \"kubernetes.io/projected/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-kube-api-access-dr4lr\") pod \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.067395 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-config-volume\") pod \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.067449 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-secret-volume\") pod \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\" (UID: \"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe\") " Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.068627 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-config-volume" (OuterVolumeSpecName: "config-volume") pod "8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe" (UID: "8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.073180 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-kube-api-access-dr4lr" (OuterVolumeSpecName: "kube-api-access-dr4lr") pod "8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe" (UID: "8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe"). InnerVolumeSpecName "kube-api-access-dr4lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.073459 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe" (UID: "8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.168893 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.168954 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr4lr\" (UniqueName: \"kubernetes.io/projected/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-kube-api-access-dr4lr\") on node \"crc\" DevicePath \"\"" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.168965 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.573706 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" event={"ID":"8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe","Type":"ContainerDied","Data":"0d565f066ee5943b8b53e69926bacd8559c64850b35827bf1264f472aff33109"} Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.573843 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.573861 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d565f066ee5943b8b53e69926bacd8559c64850b35827bf1264f472aff33109" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.592802 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:15:03 crc kubenswrapper[4946]: E1203 08:15:03.593313 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.981963 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf"] Dec 03 08:15:03 crc kubenswrapper[4946]: I1203 08:15:03.988634 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412450-wxddf"] Dec 03 08:15:05 crc kubenswrapper[4946]: I1203 08:15:05.610666 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0069d43-de07-40a9-84aa-11d303fb01b0" path="/var/lib/kubelet/pods/d0069d43-de07-40a9-84aa-11d303fb01b0/volumes" Dec 03 08:15:14 crc kubenswrapper[4946]: I1203 08:15:14.317051 4946 scope.go:117] "RemoveContainer" containerID="02f75138b6c78b38777921bacfcc64dd4078032bd386a56d9d7894bd8f2e10c3" Dec 03 08:15:18 crc kubenswrapper[4946]: I1203 08:15:18.592683 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:15:18 crc kubenswrapper[4946]: E1203 08:15:18.593449 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:15:33 crc kubenswrapper[4946]: I1203 08:15:33.593693 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:15:33 crc kubenswrapper[4946]: E1203 08:15:33.594885 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:15:47 crc kubenswrapper[4946]: I1203 08:15:47.597047 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:15:47 crc kubenswrapper[4946]: E1203 08:15:47.598151 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:16:02 crc kubenswrapper[4946]: I1203 08:16:02.593895 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:16:02 crc kubenswrapper[4946]: E1203 08:16:02.595262 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:16:13 crc kubenswrapper[4946]: I1203 08:16:13.592481 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:16:13 crc kubenswrapper[4946]: E1203 08:16:13.594329 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:16:27 crc kubenswrapper[4946]: I1203 08:16:27.601256 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:16:27 crc kubenswrapper[4946]: E1203 08:16:27.602484 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:16:41 crc kubenswrapper[4946]: I1203 08:16:41.592676 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:16:41 crc kubenswrapper[4946]: E1203 08:16:41.593851 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:16:54 crc kubenswrapper[4946]: I1203 08:16:54.592401 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:16:54 crc kubenswrapper[4946]: E1203 08:16:54.593235 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:17:08 crc kubenswrapper[4946]: I1203 08:17:08.593267 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:17:08 crc kubenswrapper[4946]: E1203 08:17:08.594385 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:17:22 crc kubenswrapper[4946]: I1203 08:17:22.592898 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:17:22 crc kubenswrapper[4946]: E1203 08:17:22.593986 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:17:35 crc kubenswrapper[4946]: I1203 08:17:35.592574 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:17:35 crc kubenswrapper[4946]: E1203 08:17:35.593460 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:17:49 crc kubenswrapper[4946]: I1203 08:17:49.593729 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:17:49 crc kubenswrapper[4946]: E1203 08:17:49.594702 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:18:01 crc kubenswrapper[4946]: I1203 08:18:01.593726 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:18:01 crc kubenswrapper[4946]: E1203 08:18:01.594994 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.286624 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-44cbf"] Dec 03 08:18:13 crc kubenswrapper[4946]: E1203 08:18:13.287347 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe" containerName="collect-profiles" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.287360 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe" containerName="collect-profiles" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.287501 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe" containerName="collect-profiles" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.288506 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.321681 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-44cbf"] Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.348501 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-utilities\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.348573 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-catalog-content\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.348697 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbptq\" (UniqueName: \"kubernetes.io/projected/ea9890ae-4668-4101-b2bd-b55ce59deab0-kube-api-access-mbptq\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.450213 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-utilities\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.450276 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-catalog-content\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.450369 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbptq\" (UniqueName: \"kubernetes.io/projected/ea9890ae-4668-4101-b2bd-b55ce59deab0-kube-api-access-mbptq\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.450910 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-utilities\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.451121 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-catalog-content\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.476679 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbptq\" (UniqueName: \"kubernetes.io/projected/ea9890ae-4668-4101-b2bd-b55ce59deab0-kube-api-access-mbptq\") pod \"community-operators-44cbf\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.610141 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:13 crc kubenswrapper[4946]: I1203 08:18:13.969293 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-44cbf"] Dec 03 08:18:14 crc kubenswrapper[4946]: I1203 08:18:14.408840 4946 generic.go:334] "Generic (PLEG): container finished" podID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerID="909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e" exitCode=0 Dec 03 08:18:14 crc kubenswrapper[4946]: I1203 08:18:14.408882 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44cbf" event={"ID":"ea9890ae-4668-4101-b2bd-b55ce59deab0","Type":"ContainerDied","Data":"909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e"} Dec 03 08:18:14 crc kubenswrapper[4946]: I1203 08:18:14.408906 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44cbf" event={"ID":"ea9890ae-4668-4101-b2bd-b55ce59deab0","Type":"ContainerStarted","Data":"7058c2c161db323d3efa9bf72d014b1de37113501d924a683de4cbb2884b1ce2"} Dec 03 08:18:14 crc kubenswrapper[4946]: I1203 08:18:14.410998 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:18:14 crc kubenswrapper[4946]: I1203 08:18:14.592891 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:18:14 crc kubenswrapper[4946]: E1203 08:18:14.593242 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:18:15 crc kubenswrapper[4946]: I1203 08:18:15.418245 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44cbf" event={"ID":"ea9890ae-4668-4101-b2bd-b55ce59deab0","Type":"ContainerStarted","Data":"cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c"} Dec 03 08:18:16 crc kubenswrapper[4946]: I1203 08:18:16.426080 4946 generic.go:334] "Generic (PLEG): container finished" podID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerID="cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c" exitCode=0 Dec 03 08:18:16 crc kubenswrapper[4946]: I1203 08:18:16.426169 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44cbf" event={"ID":"ea9890ae-4668-4101-b2bd-b55ce59deab0","Type":"ContainerDied","Data":"cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c"} Dec 03 08:18:17 crc kubenswrapper[4946]: I1203 08:18:17.435529 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44cbf" event={"ID":"ea9890ae-4668-4101-b2bd-b55ce59deab0","Type":"ContainerStarted","Data":"ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3"} Dec 03 08:18:17 crc kubenswrapper[4946]: I1203 08:18:17.464734 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-44cbf" podStartSLOduration=1.924768813 podStartE2EDuration="4.464696864s" podCreationTimestamp="2025-12-03 08:18:13 +0000 UTC" firstStartedPulling="2025-12-03 08:18:14.410721556 +0000 UTC m=+5287.207411655" lastFinishedPulling="2025-12-03 08:18:16.950649597 +0000 UTC m=+5289.747339706" observedRunningTime="2025-12-03 08:18:17.460437901 +0000 UTC m=+5290.257128010" watchObservedRunningTime="2025-12-03 08:18:17.464696864 +0000 UTC m=+5290.261386973" Dec 03 08:18:23 crc kubenswrapper[4946]: I1203 08:18:23.617203 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:23 crc kubenswrapper[4946]: I1203 08:18:23.618108 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:23 crc kubenswrapper[4946]: I1203 08:18:23.692197 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:24 crc kubenswrapper[4946]: I1203 08:18:24.558558 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:24 crc kubenswrapper[4946]: I1203 08:18:24.607376 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-44cbf"] Dec 03 08:18:26 crc kubenswrapper[4946]: I1203 08:18:26.523091 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-44cbf" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerName="registry-server" containerID="cri-o://ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3" gracePeriod=2 Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.073469 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.187788 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-catalog-content\") pod \"ea9890ae-4668-4101-b2bd-b55ce59deab0\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.187898 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbptq\" (UniqueName: \"kubernetes.io/projected/ea9890ae-4668-4101-b2bd-b55ce59deab0-kube-api-access-mbptq\") pod \"ea9890ae-4668-4101-b2bd-b55ce59deab0\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.187958 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-utilities\") pod \"ea9890ae-4668-4101-b2bd-b55ce59deab0\" (UID: \"ea9890ae-4668-4101-b2bd-b55ce59deab0\") " Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.188796 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-utilities" (OuterVolumeSpecName: "utilities") pod "ea9890ae-4668-4101-b2bd-b55ce59deab0" (UID: "ea9890ae-4668-4101-b2bd-b55ce59deab0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.194014 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9890ae-4668-4101-b2bd-b55ce59deab0-kube-api-access-mbptq" (OuterVolumeSpecName: "kube-api-access-mbptq") pod "ea9890ae-4668-4101-b2bd-b55ce59deab0" (UID: "ea9890ae-4668-4101-b2bd-b55ce59deab0"). InnerVolumeSpecName "kube-api-access-mbptq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.244208 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea9890ae-4668-4101-b2bd-b55ce59deab0" (UID: "ea9890ae-4668-4101-b2bd-b55ce59deab0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.289674 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbptq\" (UniqueName: \"kubernetes.io/projected/ea9890ae-4668-4101-b2bd-b55ce59deab0-kube-api-access-mbptq\") on node \"crc\" DevicePath \"\"" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.289730 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.289777 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9890ae-4668-4101-b2bd-b55ce59deab0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.544806 4946 generic.go:334] "Generic (PLEG): container finished" podID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerID="ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3" exitCode=0 Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.544860 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44cbf" event={"ID":"ea9890ae-4668-4101-b2bd-b55ce59deab0","Type":"ContainerDied","Data":"ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3"} Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.544895 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44cbf" event={"ID":"ea9890ae-4668-4101-b2bd-b55ce59deab0","Type":"ContainerDied","Data":"7058c2c161db323d3efa9bf72d014b1de37113501d924a683de4cbb2884b1ce2"} Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.544917 4946 scope.go:117] "RemoveContainer" containerID="ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.544916 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-44cbf" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.587536 4946 scope.go:117] "RemoveContainer" containerID="cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.589394 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-44cbf"] Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.592827 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:18:28 crc kubenswrapper[4946]: E1203 08:18:28.593195 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.596466 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-44cbf"] Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.618179 4946 scope.go:117] "RemoveContainer" containerID="909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.661469 4946 scope.go:117] "RemoveContainer" containerID="ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3" Dec 03 08:18:28 crc kubenswrapper[4946]: E1203 08:18:28.661981 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3\": container with ID starting with ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3 not found: ID does not exist" containerID="ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.662074 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3"} err="failed to get container status \"ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3\": rpc error: code = NotFound desc = could not find container \"ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3\": container with ID starting with ba34b79c01416af8ee1fbc3a5f36464890936508090a1788e1647957de3bdcb3 not found: ID does not exist" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.662112 4946 scope.go:117] "RemoveContainer" containerID="cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c" Dec 03 08:18:28 crc kubenswrapper[4946]: E1203 08:18:28.662460 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c\": container with ID starting with cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c not found: ID does not exist" containerID="cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.662491 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c"} err="failed to get container status \"cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c\": rpc error: code = NotFound desc = could not find container \"cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c\": container with ID starting with cf8d15f4a264aebd0957f934c79b0271d08db0564a7b9c18f85b0897fa90422c not found: ID does not exist" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.662511 4946 scope.go:117] "RemoveContainer" containerID="909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e" Dec 03 08:18:28 crc kubenswrapper[4946]: E1203 08:18:28.662990 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e\": container with ID starting with 909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e not found: ID does not exist" containerID="909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e" Dec 03 08:18:28 crc kubenswrapper[4946]: I1203 08:18:28.663045 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e"} err="failed to get container status \"909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e\": rpc error: code = NotFound desc = could not find container \"909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e\": container with ID starting with 909b857ef243ebed8c40d3de182c95829ce6daa31483bac152c34630d6acc23e not found: ID does not exist" Dec 03 08:18:29 crc kubenswrapper[4946]: I1203 08:18:29.606783 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" path="/var/lib/kubelet/pods/ea9890ae-4668-4101-b2bd-b55ce59deab0/volumes" Dec 03 08:18:43 crc kubenswrapper[4946]: I1203 08:18:43.593169 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:18:43 crc kubenswrapper[4946]: E1203 08:18:43.594104 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:18:55 crc kubenswrapper[4946]: I1203 08:18:55.593674 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:18:55 crc kubenswrapper[4946]: E1203 08:18:55.594874 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:19:07 crc kubenswrapper[4946]: I1203 08:19:07.602352 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:19:07 crc kubenswrapper[4946]: E1203 08:19:07.603647 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:19:20 crc kubenswrapper[4946]: I1203 08:19:20.592847 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:19:20 crc kubenswrapper[4946]: E1203 08:19:20.593664 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:19:33 crc kubenswrapper[4946]: I1203 08:19:33.593017 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:19:34 crc kubenswrapper[4946]: I1203 08:19:34.126467 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"0823ead2234ef09d244d099f7178a9d351476d79b6fd1358b7868bbfc4aba4cb"} Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.354802 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ltnd6"] Dec 03 08:20:19 crc kubenswrapper[4946]: E1203 08:20:19.356051 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerName="extract-utilities" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.356079 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerName="extract-utilities" Dec 03 08:20:19 crc kubenswrapper[4946]: E1203 08:20:19.356117 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerName="extract-content" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.356130 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerName="extract-content" Dec 03 08:20:19 crc kubenswrapper[4946]: E1203 08:20:19.356158 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerName="registry-server" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.356171 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerName="registry-server" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.356404 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea9890ae-4668-4101-b2bd-b55ce59deab0" containerName="registry-server" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.358126 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.362522 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ltnd6"] Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.512449 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-catalog-content\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.512503 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8wmw\" (UniqueName: \"kubernetes.io/projected/6dccf6c1-8383-4945-add3-4fa9d98b7231-kube-api-access-g8wmw\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.512582 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-utilities\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.614062 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-catalog-content\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.614111 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8wmw\" (UniqueName: \"kubernetes.io/projected/6dccf6c1-8383-4945-add3-4fa9d98b7231-kube-api-access-g8wmw\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.614130 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-utilities\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.614659 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-utilities\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.614662 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-catalog-content\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.639064 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8wmw\" (UniqueName: \"kubernetes.io/projected/6dccf6c1-8383-4945-add3-4fa9d98b7231-kube-api-access-g8wmw\") pod \"redhat-operators-ltnd6\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.691060 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:19 crc kubenswrapper[4946]: I1203 08:20:19.903320 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ltnd6"] Dec 03 08:20:20 crc kubenswrapper[4946]: I1203 08:20:20.604579 4946 generic.go:334] "Generic (PLEG): container finished" podID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerID="ca907d69fee116cc6b816dbc1f01d6c47ba774dce6a6d6cbca92c452d3a8f92b" exitCode=0 Dec 03 08:20:20 crc kubenswrapper[4946]: I1203 08:20:20.604642 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltnd6" event={"ID":"6dccf6c1-8383-4945-add3-4fa9d98b7231","Type":"ContainerDied","Data":"ca907d69fee116cc6b816dbc1f01d6c47ba774dce6a6d6cbca92c452d3a8f92b"} Dec 03 08:20:20 crc kubenswrapper[4946]: I1203 08:20:20.605889 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltnd6" event={"ID":"6dccf6c1-8383-4945-add3-4fa9d98b7231","Type":"ContainerStarted","Data":"bf25b1dfa4666fac1abc43b8f029d505eca619bb7d0350e4cf0e2d636859edd9"} Dec 03 08:20:22 crc kubenswrapper[4946]: I1203 08:20:22.625477 4946 generic.go:334] "Generic (PLEG): container finished" podID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerID="85bc65700dead49aa8acc9f7006dd798a4fd6226323afc6448eb7503d7e299b0" exitCode=0 Dec 03 08:20:22 crc kubenswrapper[4946]: I1203 08:20:22.625560 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltnd6" event={"ID":"6dccf6c1-8383-4945-add3-4fa9d98b7231","Type":"ContainerDied","Data":"85bc65700dead49aa8acc9f7006dd798a4fd6226323afc6448eb7503d7e299b0"} Dec 03 08:20:23 crc kubenswrapper[4946]: I1203 08:20:23.637654 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltnd6" event={"ID":"6dccf6c1-8383-4945-add3-4fa9d98b7231","Type":"ContainerStarted","Data":"784f03266cf21819d427f2b4bf0a296f07fc74a4aae27ea7d0f3df8bf41e278d"} Dec 03 08:20:23 crc kubenswrapper[4946]: I1203 08:20:23.659841 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ltnd6" podStartSLOduration=2.131869066 podStartE2EDuration="4.65980783s" podCreationTimestamp="2025-12-03 08:20:19 +0000 UTC" firstStartedPulling="2025-12-03 08:20:20.60767989 +0000 UTC m=+5413.404370009" lastFinishedPulling="2025-12-03 08:20:23.135618654 +0000 UTC m=+5415.932308773" observedRunningTime="2025-12-03 08:20:23.656366668 +0000 UTC m=+5416.453056787" watchObservedRunningTime="2025-12-03 08:20:23.65980783 +0000 UTC m=+5416.456497949" Dec 03 08:20:29 crc kubenswrapper[4946]: I1203 08:20:29.691920 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:29 crc kubenswrapper[4946]: I1203 08:20:29.692602 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:30 crc kubenswrapper[4946]: I1203 08:20:30.784163 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ltnd6" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="registry-server" probeResult="failure" output=< Dec 03 08:20:30 crc kubenswrapper[4946]: timeout: failed to connect service ":50051" within 1s Dec 03 08:20:30 crc kubenswrapper[4946]: > Dec 03 08:20:39 crc kubenswrapper[4946]: I1203 08:20:39.737211 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:39 crc kubenswrapper[4946]: I1203 08:20:39.783411 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:39 crc kubenswrapper[4946]: I1203 08:20:39.970021 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ltnd6"] Dec 03 08:20:40 crc kubenswrapper[4946]: I1203 08:20:40.781421 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ltnd6" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="registry-server" containerID="cri-o://784f03266cf21819d427f2b4bf0a296f07fc74a4aae27ea7d0f3df8bf41e278d" gracePeriod=2 Dec 03 08:20:43 crc kubenswrapper[4946]: I1203 08:20:43.807223 4946 generic.go:334] "Generic (PLEG): container finished" podID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerID="784f03266cf21819d427f2b4bf0a296f07fc74a4aae27ea7d0f3df8bf41e278d" exitCode=0 Dec 03 08:20:43 crc kubenswrapper[4946]: I1203 08:20:43.807708 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltnd6" event={"ID":"6dccf6c1-8383-4945-add3-4fa9d98b7231","Type":"ContainerDied","Data":"784f03266cf21819d427f2b4bf0a296f07fc74a4aae27ea7d0f3df8bf41e278d"} Dec 03 08:20:43 crc kubenswrapper[4946]: I1203 08:20:43.880714 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:43 crc kubenswrapper[4946]: I1203 08:20:43.953909 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-utilities\") pod \"6dccf6c1-8383-4945-add3-4fa9d98b7231\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " Dec 03 08:20:43 crc kubenswrapper[4946]: I1203 08:20:43.954102 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-catalog-content\") pod \"6dccf6c1-8383-4945-add3-4fa9d98b7231\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " Dec 03 08:20:43 crc kubenswrapper[4946]: I1203 08:20:43.954156 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8wmw\" (UniqueName: \"kubernetes.io/projected/6dccf6c1-8383-4945-add3-4fa9d98b7231-kube-api-access-g8wmw\") pod \"6dccf6c1-8383-4945-add3-4fa9d98b7231\" (UID: \"6dccf6c1-8383-4945-add3-4fa9d98b7231\") " Dec 03 08:20:43 crc kubenswrapper[4946]: I1203 08:20:43.955355 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-utilities" (OuterVolumeSpecName: "utilities") pod "6dccf6c1-8383-4945-add3-4fa9d98b7231" (UID: "6dccf6c1-8383-4945-add3-4fa9d98b7231"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:20:43 crc kubenswrapper[4946]: I1203 08:20:43.966190 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dccf6c1-8383-4945-add3-4fa9d98b7231-kube-api-access-g8wmw" (OuterVolumeSpecName: "kube-api-access-g8wmw") pod "6dccf6c1-8383-4945-add3-4fa9d98b7231" (UID: "6dccf6c1-8383-4945-add3-4fa9d98b7231"). InnerVolumeSpecName "kube-api-access-g8wmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.055923 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8wmw\" (UniqueName: \"kubernetes.io/projected/6dccf6c1-8383-4945-add3-4fa9d98b7231-kube-api-access-g8wmw\") on node \"crc\" DevicePath \"\"" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.055972 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.093047 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6dccf6c1-8383-4945-add3-4fa9d98b7231" (UID: "6dccf6c1-8383-4945-add3-4fa9d98b7231"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.158606 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dccf6c1-8383-4945-add3-4fa9d98b7231-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.822623 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltnd6" event={"ID":"6dccf6c1-8383-4945-add3-4fa9d98b7231","Type":"ContainerDied","Data":"bf25b1dfa4666fac1abc43b8f029d505eca619bb7d0350e4cf0e2d636859edd9"} Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.822680 4946 scope.go:117] "RemoveContainer" containerID="784f03266cf21819d427f2b4bf0a296f07fc74a4aae27ea7d0f3df8bf41e278d" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.822727 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltnd6" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.858724 4946 scope.go:117] "RemoveContainer" containerID="85bc65700dead49aa8acc9f7006dd798a4fd6226323afc6448eb7503d7e299b0" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.883545 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ltnd6"] Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.893781 4946 scope.go:117] "RemoveContainer" containerID="ca907d69fee116cc6b816dbc1f01d6c47ba774dce6a6d6cbca92c452d3a8f92b" Dec 03 08:20:44 crc kubenswrapper[4946]: I1203 08:20:44.898671 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ltnd6"] Dec 03 08:20:45 crc kubenswrapper[4946]: I1203 08:20:45.606017 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" path="/var/lib/kubelet/pods/6dccf6c1-8383-4945-add3-4fa9d98b7231/volumes" Dec 03 08:21:53 crc kubenswrapper[4946]: I1203 08:21:53.039910 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:21:53 crc kubenswrapper[4946]: I1203 08:21:53.040627 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:22:23 crc kubenswrapper[4946]: I1203 08:22:23.040170 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:22:23 crc kubenswrapper[4946]: I1203 08:22:23.040790 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:22:53 crc kubenswrapper[4946]: I1203 08:22:53.039289 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:22:53 crc kubenswrapper[4946]: I1203 08:22:53.040071 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:22:53 crc kubenswrapper[4946]: I1203 08:22:53.040151 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:22:53 crc kubenswrapper[4946]: I1203 08:22:53.040983 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0823ead2234ef09d244d099f7178a9d351476d79b6fd1358b7868bbfc4aba4cb"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:22:53 crc kubenswrapper[4946]: I1203 08:22:53.041068 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://0823ead2234ef09d244d099f7178a9d351476d79b6fd1358b7868bbfc4aba4cb" gracePeriod=600 Dec 03 08:22:54 crc kubenswrapper[4946]: I1203 08:22:54.018027 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="0823ead2234ef09d244d099f7178a9d351476d79b6fd1358b7868bbfc4aba4cb" exitCode=0 Dec 03 08:22:54 crc kubenswrapper[4946]: I1203 08:22:54.018140 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"0823ead2234ef09d244d099f7178a9d351476d79b6fd1358b7868bbfc4aba4cb"} Dec 03 08:22:54 crc kubenswrapper[4946]: I1203 08:22:54.018523 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27"} Dec 03 08:22:54 crc kubenswrapper[4946]: I1203 08:22:54.018550 4946 scope.go:117] "RemoveContainer" containerID="ec13bfea07b54aaffa8404a1def8981f42f5c8324db743ec6a4d3e3889a3a06b" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.288731 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p6jjm"] Dec 03 08:23:51 crc kubenswrapper[4946]: E1203 08:23:51.289639 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="registry-server" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.289656 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="registry-server" Dec 03 08:23:51 crc kubenswrapper[4946]: E1203 08:23:51.289671 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="extract-utilities" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.289678 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="extract-utilities" Dec 03 08:23:51 crc kubenswrapper[4946]: E1203 08:23:51.291406 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="extract-content" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.291428 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="extract-content" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.291701 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dccf6c1-8383-4945-add3-4fa9d98b7231" containerName="registry-server" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.293095 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.310566 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6jjm"] Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.484917 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-utilities\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.484988 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwfmq\" (UniqueName: \"kubernetes.io/projected/48bb6a94-75f5-4696-baf9-245a83f34d47-kube-api-access-nwfmq\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.485160 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-catalog-content\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.586278 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-utilities\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.586351 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwfmq\" (UniqueName: \"kubernetes.io/projected/48bb6a94-75f5-4696-baf9-245a83f34d47-kube-api-access-nwfmq\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.586394 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-catalog-content\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.587000 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-utilities\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.587023 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-catalog-content\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.607967 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwfmq\" (UniqueName: \"kubernetes.io/projected/48bb6a94-75f5-4696-baf9-245a83f34d47-kube-api-access-nwfmq\") pod \"redhat-marketplace-p6jjm\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:51 crc kubenswrapper[4946]: I1203 08:23:51.653687 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:23:52 crc kubenswrapper[4946]: I1203 08:23:52.079718 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6jjm"] Dec 03 08:23:52 crc kubenswrapper[4946]: I1203 08:23:52.568395 4946 generic.go:334] "Generic (PLEG): container finished" podID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerID="89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b" exitCode=0 Dec 03 08:23:52 crc kubenswrapper[4946]: I1203 08:23:52.568458 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6jjm" event={"ID":"48bb6a94-75f5-4696-baf9-245a83f34d47","Type":"ContainerDied","Data":"89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b"} Dec 03 08:23:52 crc kubenswrapper[4946]: I1203 08:23:52.568733 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6jjm" event={"ID":"48bb6a94-75f5-4696-baf9-245a83f34d47","Type":"ContainerStarted","Data":"e33bf7fee651b552de6a987e63727febd4b9707e3e7f11c5629a1ea3e30a6504"} Dec 03 08:23:52 crc kubenswrapper[4946]: I1203 08:23:52.570344 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.576577 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6jjm" event={"ID":"48bb6a94-75f5-4696-baf9-245a83f34d47","Type":"ContainerStarted","Data":"959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681"} Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.685526 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6wtft"] Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.688333 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.697203 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6wtft"] Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.818079 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntvzp\" (UniqueName: \"kubernetes.io/projected/c293f9d5-f4b3-4299-aeb1-cd24b34af626-kube-api-access-ntvzp\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.818392 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-utilities\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.818560 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-catalog-content\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.920321 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntvzp\" (UniqueName: \"kubernetes.io/projected/c293f9d5-f4b3-4299-aeb1-cd24b34af626-kube-api-access-ntvzp\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.920370 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-utilities\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.920438 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-catalog-content\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.920981 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-catalog-content\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.921134 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-utilities\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:53 crc kubenswrapper[4946]: I1203 08:23:53.939419 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntvzp\" (UniqueName: \"kubernetes.io/projected/c293f9d5-f4b3-4299-aeb1-cd24b34af626-kube-api-access-ntvzp\") pod \"certified-operators-6wtft\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:54 crc kubenswrapper[4946]: I1203 08:23:54.009036 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:23:54 crc kubenswrapper[4946]: I1203 08:23:54.432239 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6wtft"] Dec 03 08:23:54 crc kubenswrapper[4946]: W1203 08:23:54.436352 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc293f9d5_f4b3_4299_aeb1_cd24b34af626.slice/crio-aec693d04ed6154c023c0f414f2f20d1e9938e5963ab81e7fe793de2b01c80ef WatchSource:0}: Error finding container aec693d04ed6154c023c0f414f2f20d1e9938e5963ab81e7fe793de2b01c80ef: Status 404 returned error can't find the container with id aec693d04ed6154c023c0f414f2f20d1e9938e5963ab81e7fe793de2b01c80ef Dec 03 08:23:54 crc kubenswrapper[4946]: I1203 08:23:54.591475 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wtft" event={"ID":"c293f9d5-f4b3-4299-aeb1-cd24b34af626","Type":"ContainerStarted","Data":"640aebed629735b4dbfa4755114fb5f43979db0229940b2527393e1dc8f8cc72"} Dec 03 08:23:54 crc kubenswrapper[4946]: I1203 08:23:54.591579 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wtft" event={"ID":"c293f9d5-f4b3-4299-aeb1-cd24b34af626","Type":"ContainerStarted","Data":"aec693d04ed6154c023c0f414f2f20d1e9938e5963ab81e7fe793de2b01c80ef"} Dec 03 08:23:54 crc kubenswrapper[4946]: I1203 08:23:54.597823 4946 generic.go:334] "Generic (PLEG): container finished" podID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerID="959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681" exitCode=0 Dec 03 08:23:54 crc kubenswrapper[4946]: I1203 08:23:54.597865 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6jjm" event={"ID":"48bb6a94-75f5-4696-baf9-245a83f34d47","Type":"ContainerDied","Data":"959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681"} Dec 03 08:23:55 crc kubenswrapper[4946]: I1203 08:23:55.612404 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6jjm" event={"ID":"48bb6a94-75f5-4696-baf9-245a83f34d47","Type":"ContainerStarted","Data":"f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd"} Dec 03 08:23:55 crc kubenswrapper[4946]: I1203 08:23:55.613831 4946 generic.go:334] "Generic (PLEG): container finished" podID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerID="640aebed629735b4dbfa4755114fb5f43979db0229940b2527393e1dc8f8cc72" exitCode=0 Dec 03 08:23:55 crc kubenswrapper[4946]: I1203 08:23:55.613872 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wtft" event={"ID":"c293f9d5-f4b3-4299-aeb1-cd24b34af626","Type":"ContainerDied","Data":"640aebed629735b4dbfa4755114fb5f43979db0229940b2527393e1dc8f8cc72"} Dec 03 08:23:55 crc kubenswrapper[4946]: I1203 08:23:55.634240 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p6jjm" podStartSLOduration=2.144959985 podStartE2EDuration="4.634206123s" podCreationTimestamp="2025-12-03 08:23:51 +0000 UTC" firstStartedPulling="2025-12-03 08:23:52.570005396 +0000 UTC m=+5625.366695525" lastFinishedPulling="2025-12-03 08:23:55.059251554 +0000 UTC m=+5627.855941663" observedRunningTime="2025-12-03 08:23:55.633100123 +0000 UTC m=+5628.429790232" watchObservedRunningTime="2025-12-03 08:23:55.634206123 +0000 UTC m=+5628.430896232" Dec 03 08:23:56 crc kubenswrapper[4946]: I1203 08:23:56.627796 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wtft" event={"ID":"c293f9d5-f4b3-4299-aeb1-cd24b34af626","Type":"ContainerStarted","Data":"c7021fcdf498558d8295a6f2bfd18a90cbc25098618200e2611683229bfd3c92"} Dec 03 08:23:57 crc kubenswrapper[4946]: I1203 08:23:57.637641 4946 generic.go:334] "Generic (PLEG): container finished" podID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerID="c7021fcdf498558d8295a6f2bfd18a90cbc25098618200e2611683229bfd3c92" exitCode=0 Dec 03 08:23:57 crc kubenswrapper[4946]: I1203 08:23:57.637717 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wtft" event={"ID":"c293f9d5-f4b3-4299-aeb1-cd24b34af626","Type":"ContainerDied","Data":"c7021fcdf498558d8295a6f2bfd18a90cbc25098618200e2611683229bfd3c92"} Dec 03 08:23:58 crc kubenswrapper[4946]: I1203 08:23:58.651931 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wtft" event={"ID":"c293f9d5-f4b3-4299-aeb1-cd24b34af626","Type":"ContainerStarted","Data":"bc1982dcd6f13c1010b348f4d885de380103903d57dbd928285a169486599cbd"} Dec 03 08:23:58 crc kubenswrapper[4946]: I1203 08:23:58.691593 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6wtft" podStartSLOduration=2.993230886 podStartE2EDuration="5.691570118s" podCreationTimestamp="2025-12-03 08:23:53 +0000 UTC" firstStartedPulling="2025-12-03 08:23:55.615267516 +0000 UTC m=+5628.411957635" lastFinishedPulling="2025-12-03 08:23:58.313606758 +0000 UTC m=+5631.110296867" observedRunningTime="2025-12-03 08:23:58.679453484 +0000 UTC m=+5631.476143643" watchObservedRunningTime="2025-12-03 08:23:58.691570118 +0000 UTC m=+5631.488260237" Dec 03 08:24:01 crc kubenswrapper[4946]: I1203 08:24:01.654252 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:24:01 crc kubenswrapper[4946]: I1203 08:24:01.655160 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:24:01 crc kubenswrapper[4946]: I1203 08:24:01.709679 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:24:01 crc kubenswrapper[4946]: I1203 08:24:01.785313 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:24:02 crc kubenswrapper[4946]: I1203 08:24:02.077514 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6jjm"] Dec 03 08:24:03 crc kubenswrapper[4946]: I1203 08:24:03.691620 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p6jjm" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerName="registry-server" containerID="cri-o://f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd" gracePeriod=2 Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.010238 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.010813 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.076005 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.618216 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.701642 4946 generic.go:334] "Generic (PLEG): container finished" podID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerID="f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd" exitCode=0 Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.701703 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6jjm" event={"ID":"48bb6a94-75f5-4696-baf9-245a83f34d47","Type":"ContainerDied","Data":"f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd"} Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.701847 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6jjm" event={"ID":"48bb6a94-75f5-4696-baf9-245a83f34d47","Type":"ContainerDied","Data":"e33bf7fee651b552de6a987e63727febd4b9707e3e7f11c5629a1ea3e30a6504"} Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.701891 4946 scope.go:117] "RemoveContainer" containerID="f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.701769 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p6jjm" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.729203 4946 scope.go:117] "RemoveContainer" containerID="959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.755681 4946 scope.go:117] "RemoveContainer" containerID="89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.764893 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.792550 4946 scope.go:117] "RemoveContainer" containerID="f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd" Dec 03 08:24:04 crc kubenswrapper[4946]: E1203 08:24:04.793006 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd\": container with ID starting with f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd not found: ID does not exist" containerID="f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.793051 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd"} err="failed to get container status \"f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd\": rpc error: code = NotFound desc = could not find container \"f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd\": container with ID starting with f18e9e8bb59fc9498d457fb115c62b916680a0f7e9977c8140a6137f193102cd not found: ID does not exist" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.793084 4946 scope.go:117] "RemoveContainer" containerID="959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681" Dec 03 08:24:04 crc kubenswrapper[4946]: E1203 08:24:04.793554 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681\": container with ID starting with 959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681 not found: ID does not exist" containerID="959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.793578 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681"} err="failed to get container status \"959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681\": rpc error: code = NotFound desc = could not find container \"959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681\": container with ID starting with 959fe3fa667fb52ed99a2a7f59b45caf2749baae7a76b2d4cc5bd440258a4681 not found: ID does not exist" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.793596 4946 scope.go:117] "RemoveContainer" containerID="89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b" Dec 03 08:24:04 crc kubenswrapper[4946]: E1203 08:24:04.794188 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b\": container with ID starting with 89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b not found: ID does not exist" containerID="89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.794215 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b"} err="failed to get container status \"89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b\": rpc error: code = NotFound desc = could not find container \"89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b\": container with ID starting with 89ba3f3a1a00eb1c6c9f6174ec551b1e5ac702fa6814a2100a3e03beb222854b not found: ID does not exist" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.812820 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-catalog-content\") pod \"48bb6a94-75f5-4696-baf9-245a83f34d47\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.812878 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-utilities\") pod \"48bb6a94-75f5-4696-baf9-245a83f34d47\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.812907 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwfmq\" (UniqueName: \"kubernetes.io/projected/48bb6a94-75f5-4696-baf9-245a83f34d47-kube-api-access-nwfmq\") pod \"48bb6a94-75f5-4696-baf9-245a83f34d47\" (UID: \"48bb6a94-75f5-4696-baf9-245a83f34d47\") " Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.816996 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-utilities" (OuterVolumeSpecName: "utilities") pod "48bb6a94-75f5-4696-baf9-245a83f34d47" (UID: "48bb6a94-75f5-4696-baf9-245a83f34d47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.818879 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48bb6a94-75f5-4696-baf9-245a83f34d47-kube-api-access-nwfmq" (OuterVolumeSpecName: "kube-api-access-nwfmq") pod "48bb6a94-75f5-4696-baf9-245a83f34d47" (UID: "48bb6a94-75f5-4696-baf9-245a83f34d47"). InnerVolumeSpecName "kube-api-access-nwfmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.842762 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48bb6a94-75f5-4696-baf9-245a83f34d47" (UID: "48bb6a94-75f5-4696-baf9-245a83f34d47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.914785 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.914838 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48bb6a94-75f5-4696-baf9-245a83f34d47-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:24:04 crc kubenswrapper[4946]: I1203 08:24:04.914850 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwfmq\" (UniqueName: \"kubernetes.io/projected/48bb6a94-75f5-4696-baf9-245a83f34d47-kube-api-access-nwfmq\") on node \"crc\" DevicePath \"\"" Dec 03 08:24:05 crc kubenswrapper[4946]: I1203 08:24:05.029342 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6jjm"] Dec 03 08:24:05 crc kubenswrapper[4946]: I1203 08:24:05.036551 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6jjm"] Dec 03 08:24:05 crc kubenswrapper[4946]: I1203 08:24:05.610046 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" path="/var/lib/kubelet/pods/48bb6a94-75f5-4696-baf9-245a83f34d47/volumes" Dec 03 08:24:07 crc kubenswrapper[4946]: I1203 08:24:07.084157 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6wtft"] Dec 03 08:24:07 crc kubenswrapper[4946]: I1203 08:24:07.084895 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6wtft" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerName="registry-server" containerID="cri-o://bc1982dcd6f13c1010b348f4d885de380103903d57dbd928285a169486599cbd" gracePeriod=2 Dec 03 08:24:07 crc kubenswrapper[4946]: I1203 08:24:07.741829 4946 generic.go:334] "Generic (PLEG): container finished" podID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerID="bc1982dcd6f13c1010b348f4d885de380103903d57dbd928285a169486599cbd" exitCode=0 Dec 03 08:24:07 crc kubenswrapper[4946]: I1203 08:24:07.741948 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wtft" event={"ID":"c293f9d5-f4b3-4299-aeb1-cd24b34af626","Type":"ContainerDied","Data":"bc1982dcd6f13c1010b348f4d885de380103903d57dbd928285a169486599cbd"} Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.062552 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.165083 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-utilities\") pod \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.165264 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntvzp\" (UniqueName: \"kubernetes.io/projected/c293f9d5-f4b3-4299-aeb1-cd24b34af626-kube-api-access-ntvzp\") pod \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.165353 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-catalog-content\") pod \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\" (UID: \"c293f9d5-f4b3-4299-aeb1-cd24b34af626\") " Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.166140 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-utilities" (OuterVolumeSpecName: "utilities") pod "c293f9d5-f4b3-4299-aeb1-cd24b34af626" (UID: "c293f9d5-f4b3-4299-aeb1-cd24b34af626"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.174337 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c293f9d5-f4b3-4299-aeb1-cd24b34af626-kube-api-access-ntvzp" (OuterVolumeSpecName: "kube-api-access-ntvzp") pod "c293f9d5-f4b3-4299-aeb1-cd24b34af626" (UID: "c293f9d5-f4b3-4299-aeb1-cd24b34af626"). InnerVolumeSpecName "kube-api-access-ntvzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.215993 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c293f9d5-f4b3-4299-aeb1-cd24b34af626" (UID: "c293f9d5-f4b3-4299-aeb1-cd24b34af626"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.267222 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntvzp\" (UniqueName: \"kubernetes.io/projected/c293f9d5-f4b3-4299-aeb1-cd24b34af626-kube-api-access-ntvzp\") on node \"crc\" DevicePath \"\"" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.267257 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.267267 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c293f9d5-f4b3-4299-aeb1-cd24b34af626-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.756095 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6wtft" event={"ID":"c293f9d5-f4b3-4299-aeb1-cd24b34af626","Type":"ContainerDied","Data":"aec693d04ed6154c023c0f414f2f20d1e9938e5963ab81e7fe793de2b01c80ef"} Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.756184 4946 scope.go:117] "RemoveContainer" containerID="bc1982dcd6f13c1010b348f4d885de380103903d57dbd928285a169486599cbd" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.756291 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6wtft" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.788891 4946 scope.go:117] "RemoveContainer" containerID="c7021fcdf498558d8295a6f2bfd18a90cbc25098618200e2611683229bfd3c92" Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.800970 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6wtft"] Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.807845 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6wtft"] Dec 03 08:24:08 crc kubenswrapper[4946]: I1203 08:24:08.817875 4946 scope.go:117] "RemoveContainer" containerID="640aebed629735b4dbfa4755114fb5f43979db0229940b2527393e1dc8f8cc72" Dec 03 08:24:09 crc kubenswrapper[4946]: I1203 08:24:09.603537 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" path="/var/lib/kubelet/pods/c293f9d5-f4b3-4299-aeb1-cd24b34af626/volumes" Dec 03 08:24:53 crc kubenswrapper[4946]: I1203 08:24:53.039773 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:24:53 crc kubenswrapper[4946]: I1203 08:24:53.040380 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:25:23 crc kubenswrapper[4946]: I1203 08:25:23.039524 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:25:23 crc kubenswrapper[4946]: I1203 08:25:23.040221 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.039943 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.040556 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.040625 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.041350 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.041410 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" gracePeriod=600 Dec 03 08:25:53 crc kubenswrapper[4946]: E1203 08:25:53.166648 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.666838 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" exitCode=0 Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.666937 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27"} Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.667332 4946 scope.go:117] "RemoveContainer" containerID="0823ead2234ef09d244d099f7178a9d351476d79b6fd1358b7868bbfc4aba4cb" Dec 03 08:25:53 crc kubenswrapper[4946]: I1203 08:25:53.667919 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:25:53 crc kubenswrapper[4946]: E1203 08:25:53.668156 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:26:05 crc kubenswrapper[4946]: I1203 08:26:05.593295 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:26:05 crc kubenswrapper[4946]: E1203 08:26:05.594081 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:26:18 crc kubenswrapper[4946]: I1203 08:26:18.594494 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:26:18 crc kubenswrapper[4946]: E1203 08:26:18.595764 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:26:29 crc kubenswrapper[4946]: I1203 08:26:29.592997 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:26:29 crc kubenswrapper[4946]: E1203 08:26:29.593932 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:26:42 crc kubenswrapper[4946]: I1203 08:26:42.593882 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:26:42 crc kubenswrapper[4946]: E1203 08:26:42.595310 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:26:57 crc kubenswrapper[4946]: I1203 08:26:57.602832 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:26:57 crc kubenswrapper[4946]: E1203 08:26:57.604370 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:27:08 crc kubenswrapper[4946]: I1203 08:27:08.593874 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:27:08 crc kubenswrapper[4946]: E1203 08:27:08.594848 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:27:19 crc kubenswrapper[4946]: I1203 08:27:19.592643 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:27:19 crc kubenswrapper[4946]: E1203 08:27:19.593601 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:27:33 crc kubenswrapper[4946]: I1203 08:27:33.593215 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:27:33 crc kubenswrapper[4946]: E1203 08:27:33.594067 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:27:48 crc kubenswrapper[4946]: I1203 08:27:48.592384 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:27:48 crc kubenswrapper[4946]: E1203 08:27:48.593126 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:28:00 crc kubenswrapper[4946]: I1203 08:28:00.592942 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:28:00 crc kubenswrapper[4946]: E1203 08:28:00.593828 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:28:13 crc kubenswrapper[4946]: I1203 08:28:13.593255 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:28:13 crc kubenswrapper[4946]: E1203 08:28:13.593904 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:28:24 crc kubenswrapper[4946]: I1203 08:28:24.593304 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:28:24 crc kubenswrapper[4946]: E1203 08:28:24.596775 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:28:36 crc kubenswrapper[4946]: I1203 08:28:36.593026 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:28:36 crc kubenswrapper[4946]: E1203 08:28:36.593924 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:28:51 crc kubenswrapper[4946]: I1203 08:28:51.593919 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:28:51 crc kubenswrapper[4946]: E1203 08:28:51.594995 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:29:05 crc kubenswrapper[4946]: I1203 08:29:05.593545 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:29:05 crc kubenswrapper[4946]: E1203 08:29:05.594442 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:29:16 crc kubenswrapper[4946]: I1203 08:29:16.592819 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:29:16 crc kubenswrapper[4946]: E1203 08:29:16.593775 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:29:30 crc kubenswrapper[4946]: I1203 08:29:30.592722 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:29:30 crc kubenswrapper[4946]: E1203 08:29:30.593362 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:29:42 crc kubenswrapper[4946]: I1203 08:29:42.592636 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:29:42 crc kubenswrapper[4946]: E1203 08:29:42.593542 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:29:54 crc kubenswrapper[4946]: I1203 08:29:54.592722 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:29:54 crc kubenswrapper[4946]: E1203 08:29:54.594033 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.150994 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m"] Dec 03 08:30:00 crc kubenswrapper[4946]: E1203 08:30:00.151694 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerName="extract-utilities" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.151710 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerName="extract-utilities" Dec 03 08:30:00 crc kubenswrapper[4946]: E1203 08:30:00.151791 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerName="extract-content" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.151803 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerName="extract-content" Dec 03 08:30:00 crc kubenswrapper[4946]: E1203 08:30:00.151825 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerName="registry-server" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.151836 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerName="registry-server" Dec 03 08:30:00 crc kubenswrapper[4946]: E1203 08:30:00.151853 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerName="extract-utilities" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.151861 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerName="extract-utilities" Dec 03 08:30:00 crc kubenswrapper[4946]: E1203 08:30:00.151868 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerName="registry-server" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.151875 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerName="registry-server" Dec 03 08:30:00 crc kubenswrapper[4946]: E1203 08:30:00.151890 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerName="extract-content" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.151897 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerName="extract-content" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.152057 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="48bb6a94-75f5-4696-baf9-245a83f34d47" containerName="registry-server" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.152074 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="c293f9d5-f4b3-4299-aeb1-cd24b34af626" containerName="registry-server" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.152760 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.158530 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.159466 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.167548 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m"] Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.230065 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-secret-volume\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.230202 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffgd7\" (UniqueName: \"kubernetes.io/projected/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-kube-api-access-ffgd7\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.230308 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-config-volume\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.332059 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffgd7\" (UniqueName: \"kubernetes.io/projected/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-kube-api-access-ffgd7\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.332169 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-config-volume\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.332225 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-secret-volume\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.334701 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-config-volume\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.346104 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-secret-volume\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.348942 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffgd7\" (UniqueName: \"kubernetes.io/projected/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-kube-api-access-ffgd7\") pod \"collect-profiles-29412510-5jn5m\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.524950 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:00 crc kubenswrapper[4946]: I1203 08:30:00.797781 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m"] Dec 03 08:30:01 crc kubenswrapper[4946]: I1203 08:30:01.792772 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" event={"ID":"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3","Type":"ContainerStarted","Data":"3a89af099c1e3524ebd038ec080c79c4f44a8fc51feeec17e9b63f664aa5795f"} Dec 03 08:30:01 crc kubenswrapper[4946]: I1203 08:30:01.793062 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" event={"ID":"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3","Type":"ContainerStarted","Data":"5dd7a0dc6515285644b6fb4d8c36100556786affdac11175ed8e38ce2f42d0c2"} Dec 03 08:30:01 crc kubenswrapper[4946]: I1203 08:30:01.822986 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" podStartSLOduration=1.82292404 podStartE2EDuration="1.82292404s" podCreationTimestamp="2025-12-03 08:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:30:01.813242566 +0000 UTC m=+5994.609932675" watchObservedRunningTime="2025-12-03 08:30:01.82292404 +0000 UTC m=+5994.619614159" Dec 03 08:30:02 crc kubenswrapper[4946]: I1203 08:30:02.802682 4946 generic.go:334] "Generic (PLEG): container finished" podID="5fa8e01b-aee0-426e-bca2-9a4678b0e9f3" containerID="3a89af099c1e3524ebd038ec080c79c4f44a8fc51feeec17e9b63f664aa5795f" exitCode=0 Dec 03 08:30:02 crc kubenswrapper[4946]: I1203 08:30:02.802751 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" event={"ID":"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3","Type":"ContainerDied","Data":"3a89af099c1e3524ebd038ec080c79c4f44a8fc51feeec17e9b63f664aa5795f"} Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.087575 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.186565 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffgd7\" (UniqueName: \"kubernetes.io/projected/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-kube-api-access-ffgd7\") pod \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.186704 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-secret-volume\") pod \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.186772 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-config-volume\") pod \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\" (UID: \"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3\") " Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.187686 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-config-volume" (OuterVolumeSpecName: "config-volume") pod "5fa8e01b-aee0-426e-bca2-9a4678b0e9f3" (UID: "5fa8e01b-aee0-426e-bca2-9a4678b0e9f3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.194367 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-kube-api-access-ffgd7" (OuterVolumeSpecName: "kube-api-access-ffgd7") pod "5fa8e01b-aee0-426e-bca2-9a4678b0e9f3" (UID: "5fa8e01b-aee0-426e-bca2-9a4678b0e9f3"). InnerVolumeSpecName "kube-api-access-ffgd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.195991 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5fa8e01b-aee0-426e-bca2-9a4678b0e9f3" (UID: "5fa8e01b-aee0-426e-bca2-9a4678b0e9f3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.288502 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffgd7\" (UniqueName: \"kubernetes.io/projected/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-kube-api-access-ffgd7\") on node \"crc\" DevicePath \"\"" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.288551 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.288562 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.822730 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" event={"ID":"5fa8e01b-aee0-426e-bca2-9a4678b0e9f3","Type":"ContainerDied","Data":"5dd7a0dc6515285644b6fb4d8c36100556786affdac11175ed8e38ce2f42d0c2"} Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.822826 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dd7a0dc6515285644b6fb4d8c36100556786affdac11175ed8e38ce2f42d0c2" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.822891 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m" Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.890436 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd"] Dec 03 08:30:04 crc kubenswrapper[4946]: I1203 08:30:04.895036 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412465-j59sd"] Dec 03 08:30:05 crc kubenswrapper[4946]: I1203 08:30:05.593709 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:30:05 crc kubenswrapper[4946]: E1203 08:30:05.594448 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:30:05 crc kubenswrapper[4946]: I1203 08:30:05.609232 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24c5cc9e-1a0d-4d92-9f86-61684d69f75d" path="/var/lib/kubelet/pods/24c5cc9e-1a0d-4d92-9f86-61684d69f75d/volumes" Dec 03 08:30:14 crc kubenswrapper[4946]: I1203 08:30:14.688698 4946 scope.go:117] "RemoveContainer" containerID="f39203099fbc636700244e31c85f7af3078fad13445b07d580c81ca54ef64cbe" Dec 03 08:30:18 crc kubenswrapper[4946]: I1203 08:30:18.592561 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:30:18 crc kubenswrapper[4946]: E1203 08:30:18.593358 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:30:33 crc kubenswrapper[4946]: I1203 08:30:33.592730 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:30:33 crc kubenswrapper[4946]: E1203 08:30:33.593697 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:30:47 crc kubenswrapper[4946]: I1203 08:30:47.598069 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:30:47 crc kubenswrapper[4946]: E1203 08:30:47.599035 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:31:00 crc kubenswrapper[4946]: I1203 08:31:00.593452 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:31:02 crc kubenswrapper[4946]: I1203 08:31:02.299718 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"0b26a23d2ce0998b948bddeacdb71cdca7112080981c50a95479e9b7cd8c34a5"} Dec 03 08:33:23 crc kubenswrapper[4946]: I1203 08:33:23.039801 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:33:23 crc kubenswrapper[4946]: I1203 08:33:23.040687 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:33:53 crc kubenswrapper[4946]: I1203 08:33:53.039655 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:33:53 crc kubenswrapper[4946]: I1203 08:33:53.040368 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.191629 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9b477"] Dec 03 08:34:08 crc kubenswrapper[4946]: E1203 08:34:08.207304 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa8e01b-aee0-426e-bca2-9a4678b0e9f3" containerName="collect-profiles" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.207327 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa8e01b-aee0-426e-bca2-9a4678b0e9f3" containerName="collect-profiles" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.207548 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fa8e01b-aee0-426e-bca2-9a4678b0e9f3" containerName="collect-profiles" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.208846 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9b477"] Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.208966 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.334672 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-utilities\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.334801 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-catalog-content\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.334841 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgs9r\" (UniqueName: \"kubernetes.io/projected/cef9f29a-af99-4791-b7c2-21d839514506-kube-api-access-sgs9r\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.436305 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-utilities\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.436400 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-catalog-content\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.436432 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgs9r\" (UniqueName: \"kubernetes.io/projected/cef9f29a-af99-4791-b7c2-21d839514506-kube-api-access-sgs9r\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.436876 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-utilities\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.436973 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-catalog-content\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.457307 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgs9r\" (UniqueName: \"kubernetes.io/projected/cef9f29a-af99-4791-b7c2-21d839514506-kube-api-access-sgs9r\") pod \"certified-operators-9b477\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:08 crc kubenswrapper[4946]: I1203 08:34:08.534858 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:09 crc kubenswrapper[4946]: I1203 08:34:09.013019 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9b477"] Dec 03 08:34:09 crc kubenswrapper[4946]: I1203 08:34:09.054341 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9b477" event={"ID":"cef9f29a-af99-4791-b7c2-21d839514506","Type":"ContainerStarted","Data":"b8f1cf422c5fd07096a819cc9e71feb07a1d27ef01873c57c6f71009c1ae2a7b"} Dec 03 08:34:10 crc kubenswrapper[4946]: I1203 08:34:10.068646 4946 generic.go:334] "Generic (PLEG): container finished" podID="cef9f29a-af99-4791-b7c2-21d839514506" containerID="336a0b8409c7b39653e75f14f23c6930155ccd3759e2b54211b146d5cc6c53fa" exitCode=0 Dec 03 08:34:10 crc kubenswrapper[4946]: I1203 08:34:10.068721 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9b477" event={"ID":"cef9f29a-af99-4791-b7c2-21d839514506","Type":"ContainerDied","Data":"336a0b8409c7b39653e75f14f23c6930155ccd3759e2b54211b146d5cc6c53fa"} Dec 03 08:34:10 crc kubenswrapper[4946]: I1203 08:34:10.071110 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:34:11 crc kubenswrapper[4946]: I1203 08:34:11.077492 4946 generic.go:334] "Generic (PLEG): container finished" podID="cef9f29a-af99-4791-b7c2-21d839514506" containerID="c5ac6deba58b76dc1ceb09b91b9e32d0051c6826e8f5228931c7d8a20e3dde85" exitCode=0 Dec 03 08:34:11 crc kubenswrapper[4946]: I1203 08:34:11.077626 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9b477" event={"ID":"cef9f29a-af99-4791-b7c2-21d839514506","Type":"ContainerDied","Data":"c5ac6deba58b76dc1ceb09b91b9e32d0051c6826e8f5228931c7d8a20e3dde85"} Dec 03 08:34:12 crc kubenswrapper[4946]: I1203 08:34:12.086335 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9b477" event={"ID":"cef9f29a-af99-4791-b7c2-21d839514506","Type":"ContainerStarted","Data":"d443c4f6165d6d6e1e7fbfa60f0cff49fa7aede3e013951266a1d9ba0e602c88"} Dec 03 08:34:12 crc kubenswrapper[4946]: I1203 08:34:12.111551 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9b477" podStartSLOduration=2.5096781999999997 podStartE2EDuration="4.111510677s" podCreationTimestamp="2025-12-03 08:34:08 +0000 UTC" firstStartedPulling="2025-12-03 08:34:10.070788253 +0000 UTC m=+6242.867478362" lastFinishedPulling="2025-12-03 08:34:11.6726207 +0000 UTC m=+6244.469310839" observedRunningTime="2025-12-03 08:34:12.102461117 +0000 UTC m=+6244.899151246" watchObservedRunningTime="2025-12-03 08:34:12.111510677 +0000 UTC m=+6244.908200806" Dec 03 08:34:18 crc kubenswrapper[4946]: I1203 08:34:18.535507 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:18 crc kubenswrapper[4946]: I1203 08:34:18.536046 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:18 crc kubenswrapper[4946]: I1203 08:34:18.593886 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:19 crc kubenswrapper[4946]: I1203 08:34:19.217904 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:19 crc kubenswrapper[4946]: I1203 08:34:19.269630 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9b477"] Dec 03 08:34:21 crc kubenswrapper[4946]: I1203 08:34:21.165247 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9b477" podUID="cef9f29a-af99-4791-b7c2-21d839514506" containerName="registry-server" containerID="cri-o://d443c4f6165d6d6e1e7fbfa60f0cff49fa7aede3e013951266a1d9ba0e602c88" gracePeriod=2 Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.179013 4946 generic.go:334] "Generic (PLEG): container finished" podID="cef9f29a-af99-4791-b7c2-21d839514506" containerID="d443c4f6165d6d6e1e7fbfa60f0cff49fa7aede3e013951266a1d9ba0e602c88" exitCode=0 Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.179060 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9b477" event={"ID":"cef9f29a-af99-4791-b7c2-21d839514506","Type":"ContainerDied","Data":"d443c4f6165d6d6e1e7fbfa60f0cff49fa7aede3e013951266a1d9ba0e602c88"} Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.651334 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.739213 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-utilities\") pod \"cef9f29a-af99-4791-b7c2-21d839514506\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.739292 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-catalog-content\") pod \"cef9f29a-af99-4791-b7c2-21d839514506\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.739387 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgs9r\" (UniqueName: \"kubernetes.io/projected/cef9f29a-af99-4791-b7c2-21d839514506-kube-api-access-sgs9r\") pod \"cef9f29a-af99-4791-b7c2-21d839514506\" (UID: \"cef9f29a-af99-4791-b7c2-21d839514506\") " Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.740004 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-utilities" (OuterVolumeSpecName: "utilities") pod "cef9f29a-af99-4791-b7c2-21d839514506" (UID: "cef9f29a-af99-4791-b7c2-21d839514506"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.745064 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cef9f29a-af99-4791-b7c2-21d839514506-kube-api-access-sgs9r" (OuterVolumeSpecName: "kube-api-access-sgs9r") pod "cef9f29a-af99-4791-b7c2-21d839514506" (UID: "cef9f29a-af99-4791-b7c2-21d839514506"). InnerVolumeSpecName "kube-api-access-sgs9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.786670 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cef9f29a-af99-4791-b7c2-21d839514506" (UID: "cef9f29a-af99-4791-b7c2-21d839514506"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.841115 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.841165 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef9f29a-af99-4791-b7c2-21d839514506-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:34:22 crc kubenswrapper[4946]: I1203 08:34:22.841178 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgs9r\" (UniqueName: \"kubernetes.io/projected/cef9f29a-af99-4791-b7c2-21d839514506-kube-api-access-sgs9r\") on node \"crc\" DevicePath \"\"" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.039883 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.039960 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.040015 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.040929 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b26a23d2ce0998b948bddeacdb71cdca7112080981c50a95479e9b7cd8c34a5"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.041052 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://0b26a23d2ce0998b948bddeacdb71cdca7112080981c50a95479e9b7cd8c34a5" gracePeriod=600 Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.191635 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9b477" event={"ID":"cef9f29a-af99-4791-b7c2-21d839514506","Type":"ContainerDied","Data":"b8f1cf422c5fd07096a819cc9e71feb07a1d27ef01873c57c6f71009c1ae2a7b"} Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.191723 4946 scope.go:117] "RemoveContainer" containerID="d443c4f6165d6d6e1e7fbfa60f0cff49fa7aede3e013951266a1d9ba0e602c88" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.191651 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9b477" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.195251 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="0b26a23d2ce0998b948bddeacdb71cdca7112080981c50a95479e9b7cd8c34a5" exitCode=0 Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.195325 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"0b26a23d2ce0998b948bddeacdb71cdca7112080981c50a95479e9b7cd8c34a5"} Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.222675 4946 scope.go:117] "RemoveContainer" containerID="c5ac6deba58b76dc1ceb09b91b9e32d0051c6826e8f5228931c7d8a20e3dde85" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.244637 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9b477"] Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.251084 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9b477"] Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.270404 4946 scope.go:117] "RemoveContainer" containerID="336a0b8409c7b39653e75f14f23c6930155ccd3759e2b54211b146d5cc6c53fa" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.297220 4946 scope.go:117] "RemoveContainer" containerID="85f9b2ebf914175fae638b83d5485fa42bade3e990bce0a621b33b98f1479c27" Dec 03 08:34:23 crc kubenswrapper[4946]: I1203 08:34:23.602140 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cef9f29a-af99-4791-b7c2-21d839514506" path="/var/lib/kubelet/pods/cef9f29a-af99-4791-b7c2-21d839514506/volumes" Dec 03 08:34:24 crc kubenswrapper[4946]: I1203 08:34:24.208061 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff"} Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.799885 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5lkld"] Dec 03 08:34:27 crc kubenswrapper[4946]: E1203 08:34:27.800693 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef9f29a-af99-4791-b7c2-21d839514506" containerName="extract-utilities" Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.800707 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef9f29a-af99-4791-b7c2-21d839514506" containerName="extract-utilities" Dec 03 08:34:27 crc kubenswrapper[4946]: E1203 08:34:27.800750 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef9f29a-af99-4791-b7c2-21d839514506" containerName="extract-content" Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.800757 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef9f29a-af99-4791-b7c2-21d839514506" containerName="extract-content" Dec 03 08:34:27 crc kubenswrapper[4946]: E1203 08:34:27.800769 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef9f29a-af99-4791-b7c2-21d839514506" containerName="registry-server" Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.800776 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef9f29a-af99-4791-b7c2-21d839514506" containerName="registry-server" Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.801083 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef9f29a-af99-4791-b7c2-21d839514506" containerName="registry-server" Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.802872 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.811274 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5lkld"] Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.910655 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-utilities\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.911093 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lvlt\" (UniqueName: \"kubernetes.io/projected/b9a7a473-fc9a-461f-b379-ea6756eed16e-kube-api-access-7lvlt\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:27 crc kubenswrapper[4946]: I1203 08:34:27.911128 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-catalog-content\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:28 crc kubenswrapper[4946]: I1203 08:34:28.012856 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-utilities\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:28 crc kubenswrapper[4946]: I1203 08:34:28.013527 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-utilities\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:28 crc kubenswrapper[4946]: I1203 08:34:28.013674 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lvlt\" (UniqueName: \"kubernetes.io/projected/b9a7a473-fc9a-461f-b379-ea6756eed16e-kube-api-access-7lvlt\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:28 crc kubenswrapper[4946]: I1203 08:34:28.013698 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-catalog-content\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:28 crc kubenswrapper[4946]: I1203 08:34:28.014077 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-catalog-content\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:28 crc kubenswrapper[4946]: I1203 08:34:28.035034 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lvlt\" (UniqueName: \"kubernetes.io/projected/b9a7a473-fc9a-461f-b379-ea6756eed16e-kube-api-access-7lvlt\") pod \"redhat-marketplace-5lkld\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:28 crc kubenswrapper[4946]: I1203 08:34:28.149808 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:28 crc kubenswrapper[4946]: I1203 08:34:28.569552 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5lkld"] Dec 03 08:34:29 crc kubenswrapper[4946]: I1203 08:34:29.247167 4946 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerID="1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5" exitCode=0 Dec 03 08:34:29 crc kubenswrapper[4946]: I1203 08:34:29.247223 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5lkld" event={"ID":"b9a7a473-fc9a-461f-b379-ea6756eed16e","Type":"ContainerDied","Data":"1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5"} Dec 03 08:34:29 crc kubenswrapper[4946]: I1203 08:34:29.247270 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5lkld" event={"ID":"b9a7a473-fc9a-461f-b379-ea6756eed16e","Type":"ContainerStarted","Data":"6024b8dfd4cb5a29d4957f68a5fe83aa0fb3d3f96a994de5d6e1e925b54d2e80"} Dec 03 08:34:30 crc kubenswrapper[4946]: I1203 08:34:30.258886 4946 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerID="eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811" exitCode=0 Dec 03 08:34:30 crc kubenswrapper[4946]: I1203 08:34:30.258951 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5lkld" event={"ID":"b9a7a473-fc9a-461f-b379-ea6756eed16e","Type":"ContainerDied","Data":"eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811"} Dec 03 08:34:31 crc kubenswrapper[4946]: I1203 08:34:31.268941 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5lkld" event={"ID":"b9a7a473-fc9a-461f-b379-ea6756eed16e","Type":"ContainerStarted","Data":"0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282"} Dec 03 08:34:31 crc kubenswrapper[4946]: I1203 08:34:31.305232 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5lkld" podStartSLOduration=2.877702742 podStartE2EDuration="4.30521389s" podCreationTimestamp="2025-12-03 08:34:27 +0000 UTC" firstStartedPulling="2025-12-03 08:34:29.24965844 +0000 UTC m=+6262.046348549" lastFinishedPulling="2025-12-03 08:34:30.677169588 +0000 UTC m=+6263.473859697" observedRunningTime="2025-12-03 08:34:31.297446494 +0000 UTC m=+6264.094136613" watchObservedRunningTime="2025-12-03 08:34:31.30521389 +0000 UTC m=+6264.101903999" Dec 03 08:34:38 crc kubenswrapper[4946]: I1203 08:34:38.150264 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:38 crc kubenswrapper[4946]: I1203 08:34:38.151013 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:38 crc kubenswrapper[4946]: I1203 08:34:38.203196 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:38 crc kubenswrapper[4946]: I1203 08:34:38.377702 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:39 crc kubenswrapper[4946]: I1203 08:34:39.394156 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5lkld"] Dec 03 08:34:40 crc kubenswrapper[4946]: I1203 08:34:40.336163 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5lkld" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerName="registry-server" containerID="cri-o://0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282" gracePeriod=2 Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.222629 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.305065 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lvlt\" (UniqueName: \"kubernetes.io/projected/b9a7a473-fc9a-461f-b379-ea6756eed16e-kube-api-access-7lvlt\") pod \"b9a7a473-fc9a-461f-b379-ea6756eed16e\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.305132 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-utilities\") pod \"b9a7a473-fc9a-461f-b379-ea6756eed16e\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.305234 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-catalog-content\") pod \"b9a7a473-fc9a-461f-b379-ea6756eed16e\" (UID: \"b9a7a473-fc9a-461f-b379-ea6756eed16e\") " Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.306514 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-utilities" (OuterVolumeSpecName: "utilities") pod "b9a7a473-fc9a-461f-b379-ea6756eed16e" (UID: "b9a7a473-fc9a-461f-b379-ea6756eed16e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.311079 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a7a473-fc9a-461f-b379-ea6756eed16e-kube-api-access-7lvlt" (OuterVolumeSpecName: "kube-api-access-7lvlt") pod "b9a7a473-fc9a-461f-b379-ea6756eed16e" (UID: "b9a7a473-fc9a-461f-b379-ea6756eed16e"). InnerVolumeSpecName "kube-api-access-7lvlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.331675 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9a7a473-fc9a-461f-b379-ea6756eed16e" (UID: "b9a7a473-fc9a-461f-b379-ea6756eed16e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.344018 4946 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerID="0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282" exitCode=0 Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.344062 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5lkld" event={"ID":"b9a7a473-fc9a-461f-b379-ea6756eed16e","Type":"ContainerDied","Data":"0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282"} Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.344087 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5lkld" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.344108 4946 scope.go:117] "RemoveContainer" containerID="0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.344095 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5lkld" event={"ID":"b9a7a473-fc9a-461f-b379-ea6756eed16e","Type":"ContainerDied","Data":"6024b8dfd4cb5a29d4957f68a5fe83aa0fb3d3f96a994de5d6e1e925b54d2e80"} Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.364978 4946 scope.go:117] "RemoveContainer" containerID="eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.379351 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5lkld"] Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.390185 4946 scope.go:117] "RemoveContainer" containerID="1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.392424 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5lkld"] Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.410330 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.410354 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lvlt\" (UniqueName: \"kubernetes.io/projected/b9a7a473-fc9a-461f-b379-ea6756eed16e-kube-api-access-7lvlt\") on node \"crc\" DevicePath \"\"" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.410364 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a7a473-fc9a-461f-b379-ea6756eed16e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.410574 4946 scope.go:117] "RemoveContainer" containerID="0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282" Dec 03 08:34:41 crc kubenswrapper[4946]: E1203 08:34:41.411638 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282\": container with ID starting with 0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282 not found: ID does not exist" containerID="0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.411696 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282"} err="failed to get container status \"0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282\": rpc error: code = NotFound desc = could not find container \"0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282\": container with ID starting with 0af8dbae2060cf0b3f25307b8eaef1baba46a37899d3ce889cd40dfc52bad282 not found: ID does not exist" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.411720 4946 scope.go:117] "RemoveContainer" containerID="eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811" Dec 03 08:34:41 crc kubenswrapper[4946]: E1203 08:34:41.412177 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811\": container with ID starting with eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811 not found: ID does not exist" containerID="eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.412205 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811"} err="failed to get container status \"eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811\": rpc error: code = NotFound desc = could not find container \"eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811\": container with ID starting with eab4bdfc56f9ea7ef946af353ef6c72227c4b52de0a0d5943cbdedea90a51811 not found: ID does not exist" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.412223 4946 scope.go:117] "RemoveContainer" containerID="1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5" Dec 03 08:34:41 crc kubenswrapper[4946]: E1203 08:34:41.412483 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5\": container with ID starting with 1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5 not found: ID does not exist" containerID="1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.412512 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5"} err="failed to get container status \"1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5\": rpc error: code = NotFound desc = could not find container \"1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5\": container with ID starting with 1232004fcff098ab441f3ac4a024bc5219d249224da5b86e27287f19833711c5 not found: ID does not exist" Dec 03 08:34:41 crc kubenswrapper[4946]: I1203 08:34:41.602450 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" path="/var/lib/kubelet/pods/b9a7a473-fc9a-461f-b379-ea6756eed16e/volumes" Dec 03 08:36:23 crc kubenswrapper[4946]: I1203 08:36:23.039833 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:36:23 crc kubenswrapper[4946]: I1203 08:36:23.040636 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:36:53 crc kubenswrapper[4946]: I1203 08:36:53.039482 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:36:53 crc kubenswrapper[4946]: I1203 08:36:53.040271 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.618220 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6ql2b"] Dec 03 08:36:56 crc kubenswrapper[4946]: E1203 08:36:56.623014 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerName="extract-content" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.623435 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerName="extract-content" Dec 03 08:36:56 crc kubenswrapper[4946]: E1203 08:36:56.623482 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerName="registry-server" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.623495 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerName="registry-server" Dec 03 08:36:56 crc kubenswrapper[4946]: E1203 08:36:56.623532 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerName="extract-utilities" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.623553 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerName="extract-utilities" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.624013 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a473-fc9a-461f-b379-ea6756eed16e" containerName="registry-server" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.626020 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.633865 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6ql2b"] Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.773269 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-utilities\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.773342 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnbjx\" (UniqueName: \"kubernetes.io/projected/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-kube-api-access-mnbjx\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.773369 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-catalog-content\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.874365 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-utilities\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.874434 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnbjx\" (UniqueName: \"kubernetes.io/projected/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-kube-api-access-mnbjx\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.874493 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-catalog-content\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.874970 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-utilities\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.875016 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-catalog-content\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.894589 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnbjx\" (UniqueName: \"kubernetes.io/projected/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-kube-api-access-mnbjx\") pod \"community-operators-6ql2b\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:56 crc kubenswrapper[4946]: I1203 08:36:56.956829 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:36:57 crc kubenswrapper[4946]: I1203 08:36:57.538280 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6ql2b"] Dec 03 08:36:58 crc kubenswrapper[4946]: I1203 08:36:58.464685 4946 generic.go:334] "Generic (PLEG): container finished" podID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerID="b47b0d84c65497c7b6618d3a2a86f702ba1d46625054fe9a00bec41296258642" exitCode=0 Dec 03 08:36:58 crc kubenswrapper[4946]: I1203 08:36:58.464839 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ql2b" event={"ID":"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55","Type":"ContainerDied","Data":"b47b0d84c65497c7b6618d3a2a86f702ba1d46625054fe9a00bec41296258642"} Dec 03 08:36:58 crc kubenswrapper[4946]: I1203 08:36:58.465112 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ql2b" event={"ID":"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55","Type":"ContainerStarted","Data":"8298a8bb82214e3084ed18b6ce1fc0a19a3693c6265a6424c90ebd9330848aa0"} Dec 03 08:36:59 crc kubenswrapper[4946]: I1203 08:36:59.473841 4946 generic.go:334] "Generic (PLEG): container finished" podID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerID="e37eec47b9a66ea9ae3ea32f173c25f28919b45cdacc76a2641bc04132ed092e" exitCode=0 Dec 03 08:36:59 crc kubenswrapper[4946]: I1203 08:36:59.473892 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ql2b" event={"ID":"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55","Type":"ContainerDied","Data":"e37eec47b9a66ea9ae3ea32f173c25f28919b45cdacc76a2641bc04132ed092e"} Dec 03 08:37:00 crc kubenswrapper[4946]: I1203 08:37:00.486964 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ql2b" event={"ID":"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55","Type":"ContainerStarted","Data":"84d95f73be9060285d82593295b9bd57d6ec4b80c189cf8b03753c1ee8f1e9fb"} Dec 03 08:37:00 crc kubenswrapper[4946]: I1203 08:37:00.506061 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6ql2b" podStartSLOduration=3.02615922 podStartE2EDuration="4.506012944s" podCreationTimestamp="2025-12-03 08:36:56 +0000 UTC" firstStartedPulling="2025-12-03 08:36:58.466719749 +0000 UTC m=+6411.263409898" lastFinishedPulling="2025-12-03 08:36:59.946573503 +0000 UTC m=+6412.743263622" observedRunningTime="2025-12-03 08:37:00.503480067 +0000 UTC m=+6413.300170176" watchObservedRunningTime="2025-12-03 08:37:00.506012944 +0000 UTC m=+6413.302703053" Dec 03 08:37:06 crc kubenswrapper[4946]: I1203 08:37:06.958178 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:37:06 crc kubenswrapper[4946]: I1203 08:37:06.960032 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:37:07 crc kubenswrapper[4946]: I1203 08:37:07.007603 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:37:07 crc kubenswrapper[4946]: I1203 08:37:07.582817 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:37:07 crc kubenswrapper[4946]: I1203 08:37:07.634406 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6ql2b"] Dec 03 08:37:09 crc kubenswrapper[4946]: I1203 08:37:09.552470 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6ql2b" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerName="registry-server" containerID="cri-o://84d95f73be9060285d82593295b9bd57d6ec4b80c189cf8b03753c1ee8f1e9fb" gracePeriod=2 Dec 03 08:37:10 crc kubenswrapper[4946]: I1203 08:37:10.562898 4946 generic.go:334] "Generic (PLEG): container finished" podID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerID="84d95f73be9060285d82593295b9bd57d6ec4b80c189cf8b03753c1ee8f1e9fb" exitCode=0 Dec 03 08:37:10 crc kubenswrapper[4946]: I1203 08:37:10.563006 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ql2b" event={"ID":"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55","Type":"ContainerDied","Data":"84d95f73be9060285d82593295b9bd57d6ec4b80c189cf8b03753c1ee8f1e9fb"} Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.039870 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.182313 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnbjx\" (UniqueName: \"kubernetes.io/projected/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-kube-api-access-mnbjx\") pod \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.182357 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-utilities\") pod \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.182386 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-catalog-content\") pod \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\" (UID: \"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55\") " Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.183869 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-utilities" (OuterVolumeSpecName: "utilities") pod "8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" (UID: "8d1fbb61-bf3a-4d24-8970-6911bbc2aa55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.188919 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-kube-api-access-mnbjx" (OuterVolumeSpecName: "kube-api-access-mnbjx") pod "8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" (UID: "8d1fbb61-bf3a-4d24-8970-6911bbc2aa55"). InnerVolumeSpecName "kube-api-access-mnbjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.241971 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" (UID: "8d1fbb61-bf3a-4d24-8970-6911bbc2aa55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.284459 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnbjx\" (UniqueName: \"kubernetes.io/projected/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-kube-api-access-mnbjx\") on node \"crc\" DevicePath \"\"" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.284493 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.284504 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.575935 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ql2b" event={"ID":"8d1fbb61-bf3a-4d24-8970-6911bbc2aa55","Type":"ContainerDied","Data":"8298a8bb82214e3084ed18b6ce1fc0a19a3693c6265a6424c90ebd9330848aa0"} Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.576016 4946 scope.go:117] "RemoveContainer" containerID="84d95f73be9060285d82593295b9bd57d6ec4b80c189cf8b03753c1ee8f1e9fb" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.576030 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6ql2b" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.597407 4946 scope.go:117] "RemoveContainer" containerID="e37eec47b9a66ea9ae3ea32f173c25f28919b45cdacc76a2641bc04132ed092e" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.624426 4946 scope.go:117] "RemoveContainer" containerID="b47b0d84c65497c7b6618d3a2a86f702ba1d46625054fe9a00bec41296258642" Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.637913 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6ql2b"] Dec 03 08:37:11 crc kubenswrapper[4946]: I1203 08:37:11.650185 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6ql2b"] Dec 03 08:37:13 crc kubenswrapper[4946]: I1203 08:37:13.602644 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" path="/var/lib/kubelet/pods/8d1fbb61-bf3a-4d24-8970-6911bbc2aa55/volumes" Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.039941 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.040622 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.040684 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.041634 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.041821 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" gracePeriod=600 Dec 03 08:37:23 crc kubenswrapper[4946]: E1203 08:37:23.175174 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.692178 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" exitCode=0 Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.692289 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff"} Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.692572 4946 scope.go:117] "RemoveContainer" containerID="0b26a23d2ce0998b948bddeacdb71cdca7112080981c50a95479e9b7cd8c34a5" Dec 03 08:37:23 crc kubenswrapper[4946]: I1203 08:37:23.693500 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:37:23 crc kubenswrapper[4946]: E1203 08:37:23.693883 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:37:34 crc kubenswrapper[4946]: I1203 08:37:34.593240 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:37:34 crc kubenswrapper[4946]: E1203 08:37:34.593970 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:37:45 crc kubenswrapper[4946]: I1203 08:37:45.593229 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:37:45 crc kubenswrapper[4946]: E1203 08:37:45.594121 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:38:00 crc kubenswrapper[4946]: I1203 08:38:00.593489 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:38:00 crc kubenswrapper[4946]: E1203 08:38:00.594458 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:38:12 crc kubenswrapper[4946]: I1203 08:38:12.593473 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:38:12 crc kubenswrapper[4946]: E1203 08:38:12.594627 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:38:27 crc kubenswrapper[4946]: I1203 08:38:27.602963 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:38:27 crc kubenswrapper[4946]: E1203 08:38:27.604080 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.693341 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lv8bg"] Dec 03 08:38:31 crc kubenswrapper[4946]: E1203 08:38:31.693991 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerName="extract-content" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.694002 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerName="extract-content" Dec 03 08:38:31 crc kubenswrapper[4946]: E1203 08:38:31.694028 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerName="extract-utilities" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.694034 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerName="extract-utilities" Dec 03 08:38:31 crc kubenswrapper[4946]: E1203 08:38:31.694046 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerName="registry-server" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.694053 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerName="registry-server" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.694181 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1fbb61-bf3a-4d24-8970-6911bbc2aa55" containerName="registry-server" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.695211 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.704628 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lv8bg"] Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.834338 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-utilities\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.834462 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtpds\" (UniqueName: \"kubernetes.io/projected/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-kube-api-access-dtpds\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.834508 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-catalog-content\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.936021 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-utilities\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.936339 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtpds\" (UniqueName: \"kubernetes.io/projected/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-kube-api-access-dtpds\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.936465 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-catalog-content\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.936554 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-utilities\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.936752 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-catalog-content\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:31 crc kubenswrapper[4946]: I1203 08:38:31.960910 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtpds\" (UniqueName: \"kubernetes.io/projected/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-kube-api-access-dtpds\") pod \"redhat-operators-lv8bg\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:32 crc kubenswrapper[4946]: I1203 08:38:32.017913 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:32 crc kubenswrapper[4946]: I1203 08:38:32.440819 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lv8bg"] Dec 03 08:38:33 crc kubenswrapper[4946]: I1203 08:38:33.287719 4946 generic.go:334] "Generic (PLEG): container finished" podID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerID="ba424b4a2c6a8678704d035d58a17ba6e5b565329ed14ed9cd407c03e78b4bb7" exitCode=0 Dec 03 08:38:33 crc kubenswrapper[4946]: I1203 08:38:33.287842 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lv8bg" event={"ID":"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b","Type":"ContainerDied","Data":"ba424b4a2c6a8678704d035d58a17ba6e5b565329ed14ed9cd407c03e78b4bb7"} Dec 03 08:38:33 crc kubenswrapper[4946]: I1203 08:38:33.288076 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lv8bg" event={"ID":"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b","Type":"ContainerStarted","Data":"cb158f8909405e47f5fff609fc1aab5c1c9ed15c78d03d85ffa1674a52401919"} Dec 03 08:38:34 crc kubenswrapper[4946]: I1203 08:38:34.295246 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lv8bg" event={"ID":"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b","Type":"ContainerStarted","Data":"8cb5cd40ccb0d3b680dcda6ee603a9549ca0389425785929c17d11f79661153e"} Dec 03 08:38:35 crc kubenswrapper[4946]: I1203 08:38:35.304889 4946 generic.go:334] "Generic (PLEG): container finished" podID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerID="8cb5cd40ccb0d3b680dcda6ee603a9549ca0389425785929c17d11f79661153e" exitCode=0 Dec 03 08:38:35 crc kubenswrapper[4946]: I1203 08:38:35.304946 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lv8bg" event={"ID":"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b","Type":"ContainerDied","Data":"8cb5cd40ccb0d3b680dcda6ee603a9549ca0389425785929c17d11f79661153e"} Dec 03 08:38:36 crc kubenswrapper[4946]: I1203 08:38:36.314388 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lv8bg" event={"ID":"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b","Type":"ContainerStarted","Data":"197e439b024d74c98a893f13d67c046cec81639ac21f5660fd5b5d6fb6080a55"} Dec 03 08:38:36 crc kubenswrapper[4946]: I1203 08:38:36.330731 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lv8bg" podStartSLOduration=2.792563774 podStartE2EDuration="5.330671365s" podCreationTimestamp="2025-12-03 08:38:31 +0000 UTC" firstStartedPulling="2025-12-03 08:38:33.292436703 +0000 UTC m=+6506.089126812" lastFinishedPulling="2025-12-03 08:38:35.830544294 +0000 UTC m=+6508.627234403" observedRunningTime="2025-12-03 08:38:36.330351476 +0000 UTC m=+6509.127041615" watchObservedRunningTime="2025-12-03 08:38:36.330671365 +0000 UTC m=+6509.127361474" Dec 03 08:38:39 crc kubenswrapper[4946]: I1203 08:38:39.592577 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:38:39 crc kubenswrapper[4946]: E1203 08:38:39.593252 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:38:42 crc kubenswrapper[4946]: I1203 08:38:42.018027 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:42 crc kubenswrapper[4946]: I1203 08:38:42.018460 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:42 crc kubenswrapper[4946]: I1203 08:38:42.085428 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:42 crc kubenswrapper[4946]: I1203 08:38:42.407656 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:42 crc kubenswrapper[4946]: I1203 08:38:42.465637 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lv8bg"] Dec 03 08:38:44 crc kubenswrapper[4946]: I1203 08:38:44.377055 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lv8bg" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerName="registry-server" containerID="cri-o://197e439b024d74c98a893f13d67c046cec81639ac21f5660fd5b5d6fb6080a55" gracePeriod=2 Dec 03 08:38:46 crc kubenswrapper[4946]: I1203 08:38:46.398121 4946 generic.go:334] "Generic (PLEG): container finished" podID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerID="197e439b024d74c98a893f13d67c046cec81639ac21f5660fd5b5d6fb6080a55" exitCode=0 Dec 03 08:38:46 crc kubenswrapper[4946]: I1203 08:38:46.398161 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lv8bg" event={"ID":"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b","Type":"ContainerDied","Data":"197e439b024d74c98a893f13d67c046cec81639ac21f5660fd5b5d6fb6080a55"} Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.066587 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.099151 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-utilities\") pod \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.099266 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-catalog-content\") pod \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.099366 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtpds\" (UniqueName: \"kubernetes.io/projected/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-kube-api-access-dtpds\") pod \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\" (UID: \"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b\") " Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.101475 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-utilities" (OuterVolumeSpecName: "utilities") pod "5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" (UID: "5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.110949 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-kube-api-access-dtpds" (OuterVolumeSpecName: "kube-api-access-dtpds") pod "5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" (UID: "5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b"). InnerVolumeSpecName "kube-api-access-dtpds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.200687 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.200755 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtpds\" (UniqueName: \"kubernetes.io/projected/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-kube-api-access-dtpds\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.248169 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" (UID: "5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.301769 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.410255 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lv8bg" event={"ID":"5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b","Type":"ContainerDied","Data":"cb158f8909405e47f5fff609fc1aab5c1c9ed15c78d03d85ffa1674a52401919"} Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.410330 4946 scope.go:117] "RemoveContainer" containerID="197e439b024d74c98a893f13d67c046cec81639ac21f5660fd5b5d6fb6080a55" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.410514 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lv8bg" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.431917 4946 scope.go:117] "RemoveContainer" containerID="8cb5cd40ccb0d3b680dcda6ee603a9549ca0389425785929c17d11f79661153e" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.464852 4946 scope.go:117] "RemoveContainer" containerID="ba424b4a2c6a8678704d035d58a17ba6e5b565329ed14ed9cd407c03e78b4bb7" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.480360 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lv8bg"] Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.490963 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lv8bg"] Dec 03 08:38:47 crc kubenswrapper[4946]: E1203 08:38:47.493696 4946 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c2622a5_3eee_4f7a_8e4f_fe6b566ca29b.slice/crio-cb158f8909405e47f5fff609fc1aab5c1c9ed15c78d03d85ffa1674a52401919\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c2622a5_3eee_4f7a_8e4f_fe6b566ca29b.slice\": RecentStats: unable to find data in memory cache]" Dec 03 08:38:47 crc kubenswrapper[4946]: I1203 08:38:47.604009 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" path="/var/lib/kubelet/pods/5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b/volumes" Dec 03 08:38:53 crc kubenswrapper[4946]: I1203 08:38:53.593774 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:38:53 crc kubenswrapper[4946]: E1203 08:38:53.596215 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:39:06 crc kubenswrapper[4946]: I1203 08:39:06.593925 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:39:06 crc kubenswrapper[4946]: E1203 08:39:06.595501 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:39:20 crc kubenswrapper[4946]: I1203 08:39:20.594057 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:39:20 crc kubenswrapper[4946]: E1203 08:39:20.595419 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:39:34 crc kubenswrapper[4946]: I1203 08:39:34.593183 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:39:34 crc kubenswrapper[4946]: E1203 08:39:34.595293 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:39:47 crc kubenswrapper[4946]: I1203 08:39:47.601112 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:39:47 crc kubenswrapper[4946]: E1203 08:39:47.602544 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:40:02 crc kubenswrapper[4946]: I1203 08:40:02.592868 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:40:02 crc kubenswrapper[4946]: E1203 08:40:02.595844 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:40:13 crc kubenswrapper[4946]: I1203 08:40:13.592777 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:40:13 crc kubenswrapper[4946]: E1203 08:40:13.593529 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:40:25 crc kubenswrapper[4946]: I1203 08:40:25.592657 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:40:25 crc kubenswrapper[4946]: E1203 08:40:25.593873 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:40:38 crc kubenswrapper[4946]: I1203 08:40:38.593541 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:40:38 crc kubenswrapper[4946]: E1203 08:40:38.594555 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:40:51 crc kubenswrapper[4946]: I1203 08:40:51.593602 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:40:51 crc kubenswrapper[4946]: E1203 08:40:51.594622 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:41:06 crc kubenswrapper[4946]: I1203 08:41:06.592999 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:41:06 crc kubenswrapper[4946]: E1203 08:41:06.594488 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:41:19 crc kubenswrapper[4946]: I1203 08:41:19.593476 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:41:19 crc kubenswrapper[4946]: E1203 08:41:19.594278 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:41:33 crc kubenswrapper[4946]: I1203 08:41:33.593140 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:41:33 crc kubenswrapper[4946]: E1203 08:41:33.593876 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:41:47 crc kubenswrapper[4946]: I1203 08:41:47.598045 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:41:47 crc kubenswrapper[4946]: E1203 08:41:47.598858 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:42:00 crc kubenswrapper[4946]: I1203 08:42:00.592888 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:42:00 crc kubenswrapper[4946]: E1203 08:42:00.593993 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:42:15 crc kubenswrapper[4946]: I1203 08:42:15.592992 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:42:15 crc kubenswrapper[4946]: E1203 08:42:15.593844 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:42:26 crc kubenswrapper[4946]: I1203 08:42:26.593148 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:42:27 crc kubenswrapper[4946]: I1203 08:42:27.299916 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"878aee7b8ce5cbe55128d3e3c0dca6f3787d24f759d3a27c81faad7b98362df5"} Dec 03 08:44:53 crc kubenswrapper[4946]: I1203 08:44:53.041077 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:44:53 crc kubenswrapper[4946]: I1203 08:44:53.041999 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.164971 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q"] Dec 03 08:45:00 crc kubenswrapper[4946]: E1203 08:45:00.165924 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerName="registry-server" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.165939 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerName="registry-server" Dec 03 08:45:00 crc kubenswrapper[4946]: E1203 08:45:00.165972 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerName="extract-content" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.165981 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerName="extract-content" Dec 03 08:45:00 crc kubenswrapper[4946]: E1203 08:45:00.166002 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerName="extract-utilities" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.166011 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerName="extract-utilities" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.166197 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c2622a5-3eee-4f7a-8e4f-fe6b566ca29b" containerName="registry-server" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.166804 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.171065 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.171079 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.182473 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q"] Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.191498 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19820654-8b5b-439c-a2f2-1fc4cac8828a-secret-volume\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.191562 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7txpd\" (UniqueName: \"kubernetes.io/projected/19820654-8b5b-439c-a2f2-1fc4cac8828a-kube-api-access-7txpd\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.191637 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19820654-8b5b-439c-a2f2-1fc4cac8828a-config-volume\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.293800 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19820654-8b5b-439c-a2f2-1fc4cac8828a-secret-volume\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.293987 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7txpd\" (UniqueName: \"kubernetes.io/projected/19820654-8b5b-439c-a2f2-1fc4cac8828a-kube-api-access-7txpd\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.294062 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19820654-8b5b-439c-a2f2-1fc4cac8828a-config-volume\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.295655 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19820654-8b5b-439c-a2f2-1fc4cac8828a-config-volume\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.304131 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19820654-8b5b-439c-a2f2-1fc4cac8828a-secret-volume\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.321079 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7txpd\" (UniqueName: \"kubernetes.io/projected/19820654-8b5b-439c-a2f2-1fc4cac8828a-kube-api-access-7txpd\") pod \"collect-profiles-29412525-9l89q\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.510878 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:00 crc kubenswrapper[4946]: I1203 08:45:00.972296 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q"] Dec 03 08:45:01 crc kubenswrapper[4946]: I1203 08:45:01.833920 4946 generic.go:334] "Generic (PLEG): container finished" podID="19820654-8b5b-439c-a2f2-1fc4cac8828a" containerID="d5b9d925174d0a759fb08a3e77424f3a3102a45a88d13dea9eb2bdc986b74de3" exitCode=0 Dec 03 08:45:01 crc kubenswrapper[4946]: I1203 08:45:01.835525 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" event={"ID":"19820654-8b5b-439c-a2f2-1fc4cac8828a","Type":"ContainerDied","Data":"d5b9d925174d0a759fb08a3e77424f3a3102a45a88d13dea9eb2bdc986b74de3"} Dec 03 08:45:01 crc kubenswrapper[4946]: I1203 08:45:01.835691 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" event={"ID":"19820654-8b5b-439c-a2f2-1fc4cac8828a","Type":"ContainerStarted","Data":"f774e47ea0c5756b0c87627006b59a68a062fdaad6ddf437b6f68fe9d51139bb"} Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.223181 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.344076 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19820654-8b5b-439c-a2f2-1fc4cac8828a-config-volume\") pod \"19820654-8b5b-439c-a2f2-1fc4cac8828a\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.344258 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19820654-8b5b-439c-a2f2-1fc4cac8828a-secret-volume\") pod \"19820654-8b5b-439c-a2f2-1fc4cac8828a\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.344372 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7txpd\" (UniqueName: \"kubernetes.io/projected/19820654-8b5b-439c-a2f2-1fc4cac8828a-kube-api-access-7txpd\") pod \"19820654-8b5b-439c-a2f2-1fc4cac8828a\" (UID: \"19820654-8b5b-439c-a2f2-1fc4cac8828a\") " Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.345925 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19820654-8b5b-439c-a2f2-1fc4cac8828a-config-volume" (OuterVolumeSpecName: "config-volume") pod "19820654-8b5b-439c-a2f2-1fc4cac8828a" (UID: "19820654-8b5b-439c-a2f2-1fc4cac8828a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.354856 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19820654-8b5b-439c-a2f2-1fc4cac8828a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "19820654-8b5b-439c-a2f2-1fc4cac8828a" (UID: "19820654-8b5b-439c-a2f2-1fc4cac8828a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.355019 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19820654-8b5b-439c-a2f2-1fc4cac8828a-kube-api-access-7txpd" (OuterVolumeSpecName: "kube-api-access-7txpd") pod "19820654-8b5b-439c-a2f2-1fc4cac8828a" (UID: "19820654-8b5b-439c-a2f2-1fc4cac8828a"). InnerVolumeSpecName "kube-api-access-7txpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.446270 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19820654-8b5b-439c-a2f2-1fc4cac8828a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.446323 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19820654-8b5b-439c-a2f2-1fc4cac8828a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.446347 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7txpd\" (UniqueName: \"kubernetes.io/projected/19820654-8b5b-439c-a2f2-1fc4cac8828a-kube-api-access-7txpd\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.857680 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" event={"ID":"19820654-8b5b-439c-a2f2-1fc4cac8828a","Type":"ContainerDied","Data":"f774e47ea0c5756b0c87627006b59a68a062fdaad6ddf437b6f68fe9d51139bb"} Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.857763 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f774e47ea0c5756b0c87627006b59a68a062fdaad6ddf437b6f68fe9d51139bb" Dec 03 08:45:03 crc kubenswrapper[4946]: I1203 08:45:03.857898 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-9l89q" Dec 03 08:45:04 crc kubenswrapper[4946]: I1203 08:45:04.311799 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq"] Dec 03 08:45:04 crc kubenswrapper[4946]: I1203 08:45:04.319554 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412480-vlkgq"] Dec 03 08:45:05 crc kubenswrapper[4946]: I1203 08:45:05.605330 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbd89e07-ac18-45a6-9713-06cfc08bf40a" path="/var/lib/kubelet/pods/bbd89e07-ac18-45a6-9713-06cfc08bf40a/volumes" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.696980 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-846tl"] Dec 03 08:45:07 crc kubenswrapper[4946]: E1203 08:45:07.698100 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19820654-8b5b-439c-a2f2-1fc4cac8828a" containerName="collect-profiles" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.698127 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="19820654-8b5b-439c-a2f2-1fc4cac8828a" containerName="collect-profiles" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.698471 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="19820654-8b5b-439c-a2f2-1fc4cac8828a" containerName="collect-profiles" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.701182 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.720359 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-846tl"] Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.815033 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-utilities\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.815142 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-catalog-content\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.815209 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frrhx\" (UniqueName: \"kubernetes.io/projected/a90da788-dbde-459a-bd0f-91a525318083-kube-api-access-frrhx\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.916818 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-catalog-content\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.916884 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frrhx\" (UniqueName: \"kubernetes.io/projected/a90da788-dbde-459a-bd0f-91a525318083-kube-api-access-frrhx\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.916969 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-utilities\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.917540 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-utilities\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.917534 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-catalog-content\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:07 crc kubenswrapper[4946]: I1203 08:45:07.943428 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frrhx\" (UniqueName: \"kubernetes.io/projected/a90da788-dbde-459a-bd0f-91a525318083-kube-api-access-frrhx\") pod \"certified-operators-846tl\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:08 crc kubenswrapper[4946]: I1203 08:45:08.023468 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:08 crc kubenswrapper[4946]: I1203 08:45:08.515341 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-846tl"] Dec 03 08:45:08 crc kubenswrapper[4946]: I1203 08:45:08.905811 4946 generic.go:334] "Generic (PLEG): container finished" podID="a90da788-dbde-459a-bd0f-91a525318083" containerID="a8e2d163b20b94ad80d196778ca518271d4093eb47efe00e6eb1d7034ba6b725" exitCode=0 Dec 03 08:45:08 crc kubenswrapper[4946]: I1203 08:45:08.905877 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-846tl" event={"ID":"a90da788-dbde-459a-bd0f-91a525318083","Type":"ContainerDied","Data":"a8e2d163b20b94ad80d196778ca518271d4093eb47efe00e6eb1d7034ba6b725"} Dec 03 08:45:08 crc kubenswrapper[4946]: I1203 08:45:08.905917 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-846tl" event={"ID":"a90da788-dbde-459a-bd0f-91a525318083","Type":"ContainerStarted","Data":"360e7068093ace49db657b9d0abc166bccbda62730f2684e314c05a1b1337477"} Dec 03 08:45:08 crc kubenswrapper[4946]: I1203 08:45:08.908513 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:45:10 crc kubenswrapper[4946]: I1203 08:45:10.924849 4946 generic.go:334] "Generic (PLEG): container finished" podID="a90da788-dbde-459a-bd0f-91a525318083" containerID="c885061c8a758364e39590976da17d49379a3b63bb8c3706ade5f78dd9d95ff2" exitCode=0 Dec 03 08:45:10 crc kubenswrapper[4946]: I1203 08:45:10.925009 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-846tl" event={"ID":"a90da788-dbde-459a-bd0f-91a525318083","Type":"ContainerDied","Data":"c885061c8a758364e39590976da17d49379a3b63bb8c3706ade5f78dd9d95ff2"} Dec 03 08:45:12 crc kubenswrapper[4946]: I1203 08:45:12.945203 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-846tl" event={"ID":"a90da788-dbde-459a-bd0f-91a525318083","Type":"ContainerStarted","Data":"7a33aa28e780da3905c512acbedd50edad8e235059049ee2edf644a09782e41b"} Dec 03 08:45:12 crc kubenswrapper[4946]: I1203 08:45:12.976441 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-846tl" podStartSLOduration=3.070048318 podStartE2EDuration="5.976393271s" podCreationTimestamp="2025-12-03 08:45:07 +0000 UTC" firstStartedPulling="2025-12-03 08:45:08.908043992 +0000 UTC m=+6901.704734141" lastFinishedPulling="2025-12-03 08:45:11.814388975 +0000 UTC m=+6904.611079094" observedRunningTime="2025-12-03 08:45:12.968196304 +0000 UTC m=+6905.764886413" watchObservedRunningTime="2025-12-03 08:45:12.976393271 +0000 UTC m=+6905.773083420" Dec 03 08:45:15 crc kubenswrapper[4946]: I1203 08:45:15.061029 4946 scope.go:117] "RemoveContainer" containerID="779cad687350dfef33e55991af2adc9e22798894021254ef9f5be3a163bfced9" Dec 03 08:45:18 crc kubenswrapper[4946]: I1203 08:45:18.024567 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:18 crc kubenswrapper[4946]: I1203 08:45:18.026488 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:18 crc kubenswrapper[4946]: I1203 08:45:18.079898 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:19 crc kubenswrapper[4946]: I1203 08:45:19.075959 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:19 crc kubenswrapper[4946]: I1203 08:45:19.147623 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-846tl"] Dec 03 08:45:21 crc kubenswrapper[4946]: I1203 08:45:21.031982 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-846tl" podUID="a90da788-dbde-459a-bd0f-91a525318083" containerName="registry-server" containerID="cri-o://7a33aa28e780da3905c512acbedd50edad8e235059049ee2edf644a09782e41b" gracePeriod=2 Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.042714 4946 generic.go:334] "Generic (PLEG): container finished" podID="a90da788-dbde-459a-bd0f-91a525318083" containerID="7a33aa28e780da3905c512acbedd50edad8e235059049ee2edf644a09782e41b" exitCode=0 Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.042781 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-846tl" event={"ID":"a90da788-dbde-459a-bd0f-91a525318083","Type":"ContainerDied","Data":"7a33aa28e780da3905c512acbedd50edad8e235059049ee2edf644a09782e41b"} Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.183001 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.294119 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-utilities\") pod \"a90da788-dbde-459a-bd0f-91a525318083\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.294247 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-catalog-content\") pod \"a90da788-dbde-459a-bd0f-91a525318083\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.294360 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frrhx\" (UniqueName: \"kubernetes.io/projected/a90da788-dbde-459a-bd0f-91a525318083-kube-api-access-frrhx\") pod \"a90da788-dbde-459a-bd0f-91a525318083\" (UID: \"a90da788-dbde-459a-bd0f-91a525318083\") " Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.295000 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-utilities" (OuterVolumeSpecName: "utilities") pod "a90da788-dbde-459a-bd0f-91a525318083" (UID: "a90da788-dbde-459a-bd0f-91a525318083"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.303653 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90da788-dbde-459a-bd0f-91a525318083-kube-api-access-frrhx" (OuterVolumeSpecName: "kube-api-access-frrhx") pod "a90da788-dbde-459a-bd0f-91a525318083" (UID: "a90da788-dbde-459a-bd0f-91a525318083"). InnerVolumeSpecName "kube-api-access-frrhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.386019 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a90da788-dbde-459a-bd0f-91a525318083" (UID: "a90da788-dbde-459a-bd0f-91a525318083"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.396504 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frrhx\" (UniqueName: \"kubernetes.io/projected/a90da788-dbde-459a-bd0f-91a525318083-kube-api-access-frrhx\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.396543 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:22 crc kubenswrapper[4946]: I1203 08:45:22.396565 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90da788-dbde-459a-bd0f-91a525318083-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.039763 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.040175 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.055412 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-846tl" event={"ID":"a90da788-dbde-459a-bd0f-91a525318083","Type":"ContainerDied","Data":"360e7068093ace49db657b9d0abc166bccbda62730f2684e314c05a1b1337477"} Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.055487 4946 scope.go:117] "RemoveContainer" containerID="7a33aa28e780da3905c512acbedd50edad8e235059049ee2edf644a09782e41b" Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.055507 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-846tl" Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.086727 4946 scope.go:117] "RemoveContainer" containerID="c885061c8a758364e39590976da17d49379a3b63bb8c3706ade5f78dd9d95ff2" Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.117791 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-846tl"] Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.131360 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-846tl"] Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.145581 4946 scope.go:117] "RemoveContainer" containerID="a8e2d163b20b94ad80d196778ca518271d4093eb47efe00e6eb1d7034ba6b725" Dec 03 08:45:23 crc kubenswrapper[4946]: I1203 08:45:23.601252 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a90da788-dbde-459a-bd0f-91a525318083" path="/var/lib/kubelet/pods/a90da788-dbde-459a-bd0f-91a525318083/volumes" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.460483 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mdcvp"] Dec 03 08:45:47 crc kubenswrapper[4946]: E1203 08:45:47.461868 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90da788-dbde-459a-bd0f-91a525318083" containerName="extract-utilities" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.461892 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90da788-dbde-459a-bd0f-91a525318083" containerName="extract-utilities" Dec 03 08:45:47 crc kubenswrapper[4946]: E1203 08:45:47.461927 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90da788-dbde-459a-bd0f-91a525318083" containerName="registry-server" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.461938 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90da788-dbde-459a-bd0f-91a525318083" containerName="registry-server" Dec 03 08:45:47 crc kubenswrapper[4946]: E1203 08:45:47.461976 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90da788-dbde-459a-bd0f-91a525318083" containerName="extract-content" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.461989 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90da788-dbde-459a-bd0f-91a525318083" containerName="extract-content" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.462274 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="a90da788-dbde-459a-bd0f-91a525318083" containerName="registry-server" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.463991 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.511173 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdcvp"] Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.599013 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-catalog-content\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.599166 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-utilities\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.599246 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdgsc\" (UniqueName: \"kubernetes.io/projected/692fd2d3-4715-41af-9f10-7d11bebe6e86-kube-api-access-vdgsc\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.700864 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-catalog-content\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.701172 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-utilities\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.701208 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdgsc\" (UniqueName: \"kubernetes.io/projected/692fd2d3-4715-41af-9f10-7d11bebe6e86-kube-api-access-vdgsc\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.701506 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-catalog-content\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.701652 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-utilities\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.737387 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdgsc\" (UniqueName: \"kubernetes.io/projected/692fd2d3-4715-41af-9f10-7d11bebe6e86-kube-api-access-vdgsc\") pod \"redhat-marketplace-mdcvp\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:47 crc kubenswrapper[4946]: I1203 08:45:47.807236 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:48 crc kubenswrapper[4946]: I1203 08:45:48.430242 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdcvp"] Dec 03 08:45:48 crc kubenswrapper[4946]: W1203 08:45:48.440062 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod692fd2d3_4715_41af_9f10_7d11bebe6e86.slice/crio-fadb6101c2f28b1a821baced56c3b935d6e885e920de9937e0912d2e4044f14d WatchSource:0}: Error finding container fadb6101c2f28b1a821baced56c3b935d6e885e920de9937e0912d2e4044f14d: Status 404 returned error can't find the container with id fadb6101c2f28b1a821baced56c3b935d6e885e920de9937e0912d2e4044f14d Dec 03 08:45:49 crc kubenswrapper[4946]: I1203 08:45:49.320795 4946 generic.go:334] "Generic (PLEG): container finished" podID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerID="24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3" exitCode=0 Dec 03 08:45:49 crc kubenswrapper[4946]: I1203 08:45:49.320863 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdcvp" event={"ID":"692fd2d3-4715-41af-9f10-7d11bebe6e86","Type":"ContainerDied","Data":"24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3"} Dec 03 08:45:49 crc kubenswrapper[4946]: I1203 08:45:49.320904 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdcvp" event={"ID":"692fd2d3-4715-41af-9f10-7d11bebe6e86","Type":"ContainerStarted","Data":"fadb6101c2f28b1a821baced56c3b935d6e885e920de9937e0912d2e4044f14d"} Dec 03 08:45:50 crc kubenswrapper[4946]: I1203 08:45:50.331281 4946 generic.go:334] "Generic (PLEG): container finished" podID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerID="f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a" exitCode=0 Dec 03 08:45:50 crc kubenswrapper[4946]: I1203 08:45:50.331357 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdcvp" event={"ID":"692fd2d3-4715-41af-9f10-7d11bebe6e86","Type":"ContainerDied","Data":"f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a"} Dec 03 08:45:51 crc kubenswrapper[4946]: I1203 08:45:51.340585 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdcvp" event={"ID":"692fd2d3-4715-41af-9f10-7d11bebe6e86","Type":"ContainerStarted","Data":"79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e"} Dec 03 08:45:53 crc kubenswrapper[4946]: I1203 08:45:53.039610 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:45:53 crc kubenswrapper[4946]: I1203 08:45:53.039676 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:45:53 crc kubenswrapper[4946]: I1203 08:45:53.039715 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:45:53 crc kubenswrapper[4946]: I1203 08:45:53.040355 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"878aee7b8ce5cbe55128d3e3c0dca6f3787d24f759d3a27c81faad7b98362df5"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:45:53 crc kubenswrapper[4946]: I1203 08:45:53.040421 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://878aee7b8ce5cbe55128d3e3c0dca6f3787d24f759d3a27c81faad7b98362df5" gracePeriod=600 Dec 03 08:45:53 crc kubenswrapper[4946]: I1203 08:45:53.362317 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="878aee7b8ce5cbe55128d3e3c0dca6f3787d24f759d3a27c81faad7b98362df5" exitCode=0 Dec 03 08:45:53 crc kubenswrapper[4946]: I1203 08:45:53.362403 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"878aee7b8ce5cbe55128d3e3c0dca6f3787d24f759d3a27c81faad7b98362df5"} Dec 03 08:45:53 crc kubenswrapper[4946]: I1203 08:45:53.362645 4946 scope.go:117] "RemoveContainer" containerID="af98e688a96b31d34443d501bb551848b9c1f663b9aa575498b0bef9b3fda2ff" Dec 03 08:45:54 crc kubenswrapper[4946]: I1203 08:45:54.371369 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad"} Dec 03 08:45:54 crc kubenswrapper[4946]: I1203 08:45:54.391491 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mdcvp" podStartSLOduration=5.846566047 podStartE2EDuration="7.39145272s" podCreationTimestamp="2025-12-03 08:45:47 +0000 UTC" firstStartedPulling="2025-12-03 08:45:49.322680978 +0000 UTC m=+6942.119371097" lastFinishedPulling="2025-12-03 08:45:50.867567651 +0000 UTC m=+6943.664257770" observedRunningTime="2025-12-03 08:45:51.375492612 +0000 UTC m=+6944.172182751" watchObservedRunningTime="2025-12-03 08:45:54.39145272 +0000 UTC m=+6947.188142829" Dec 03 08:45:57 crc kubenswrapper[4946]: I1203 08:45:57.807881 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:57 crc kubenswrapper[4946]: I1203 08:45:57.811613 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:57 crc kubenswrapper[4946]: I1203 08:45:57.862228 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:58 crc kubenswrapper[4946]: I1203 08:45:58.509439 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:45:58 crc kubenswrapper[4946]: I1203 08:45:58.572961 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdcvp"] Dec 03 08:46:00 crc kubenswrapper[4946]: I1203 08:46:00.448254 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mdcvp" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerName="registry-server" containerID="cri-o://79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e" gracePeriod=2 Dec 03 08:46:00 crc kubenswrapper[4946]: I1203 08:46:00.982506 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.007518 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdgsc\" (UniqueName: \"kubernetes.io/projected/692fd2d3-4715-41af-9f10-7d11bebe6e86-kube-api-access-vdgsc\") pod \"692fd2d3-4715-41af-9f10-7d11bebe6e86\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.007677 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-utilities\") pod \"692fd2d3-4715-41af-9f10-7d11bebe6e86\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.007727 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-catalog-content\") pod \"692fd2d3-4715-41af-9f10-7d11bebe6e86\" (UID: \"692fd2d3-4715-41af-9f10-7d11bebe6e86\") " Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.012945 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-utilities" (OuterVolumeSpecName: "utilities") pod "692fd2d3-4715-41af-9f10-7d11bebe6e86" (UID: "692fd2d3-4715-41af-9f10-7d11bebe6e86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.033949 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/692fd2d3-4715-41af-9f10-7d11bebe6e86-kube-api-access-vdgsc" (OuterVolumeSpecName: "kube-api-access-vdgsc") pod "692fd2d3-4715-41af-9f10-7d11bebe6e86" (UID: "692fd2d3-4715-41af-9f10-7d11bebe6e86"). InnerVolumeSpecName "kube-api-access-vdgsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.045427 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "692fd2d3-4715-41af-9f10-7d11bebe6e86" (UID: "692fd2d3-4715-41af-9f10-7d11bebe6e86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.109419 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.109905 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdgsc\" (UniqueName: \"kubernetes.io/projected/692fd2d3-4715-41af-9f10-7d11bebe6e86-kube-api-access-vdgsc\") on node \"crc\" DevicePath \"\"" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.110128 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/692fd2d3-4715-41af-9f10-7d11bebe6e86-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.459870 4946 generic.go:334] "Generic (PLEG): container finished" podID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerID="79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e" exitCode=0 Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.459923 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdcvp" event={"ID":"692fd2d3-4715-41af-9f10-7d11bebe6e86","Type":"ContainerDied","Data":"79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e"} Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.459965 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdcvp" event={"ID":"692fd2d3-4715-41af-9f10-7d11bebe6e86","Type":"ContainerDied","Data":"fadb6101c2f28b1a821baced56c3b935d6e885e920de9937e0912d2e4044f14d"} Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.459985 4946 scope.go:117] "RemoveContainer" containerID="79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.460069 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdcvp" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.489843 4946 scope.go:117] "RemoveContainer" containerID="f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.522658 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdcvp"] Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.532256 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdcvp"] Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.547037 4946 scope.go:117] "RemoveContainer" containerID="24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.567795 4946 scope.go:117] "RemoveContainer" containerID="79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e" Dec 03 08:46:01 crc kubenswrapper[4946]: E1203 08:46:01.568356 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e\": container with ID starting with 79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e not found: ID does not exist" containerID="79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.568401 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e"} err="failed to get container status \"79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e\": rpc error: code = NotFound desc = could not find container \"79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e\": container with ID starting with 79e980f56df1c40cf86ed31f6bc83924951e9eea9c333d5d9a981c15b8db478e not found: ID does not exist" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.568426 4946 scope.go:117] "RemoveContainer" containerID="f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a" Dec 03 08:46:01 crc kubenswrapper[4946]: E1203 08:46:01.569176 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a\": container with ID starting with f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a not found: ID does not exist" containerID="f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.569203 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a"} err="failed to get container status \"f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a\": rpc error: code = NotFound desc = could not find container \"f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a\": container with ID starting with f468298970c48e09f53a2e89a3c40c5daeb5648d1e0d98aa3f2188733bf77f9a not found: ID does not exist" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.569217 4946 scope.go:117] "RemoveContainer" containerID="24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3" Dec 03 08:46:01 crc kubenswrapper[4946]: E1203 08:46:01.569538 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3\": container with ID starting with 24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3 not found: ID does not exist" containerID="24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.569594 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3"} err="failed to get container status \"24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3\": rpc error: code = NotFound desc = could not find container \"24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3\": container with ID starting with 24271fc494527f058cfce0f57664007b5c8e6b3cf61e3d1946c088ada4cba0b3 not found: ID does not exist" Dec 03 08:46:01 crc kubenswrapper[4946]: I1203 08:46:01.607918 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" path="/var/lib/kubelet/pods/692fd2d3-4715-41af-9f10-7d11bebe6e86/volumes" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.052688 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-djvq2"] Dec 03 08:47:50 crc kubenswrapper[4946]: E1203 08:47:50.053502 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerName="extract-utilities" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.053514 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerName="extract-utilities" Dec 03 08:47:50 crc kubenswrapper[4946]: E1203 08:47:50.053534 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerName="registry-server" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.053540 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerName="registry-server" Dec 03 08:47:50 crc kubenswrapper[4946]: E1203 08:47:50.053551 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerName="extract-content" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.053557 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerName="extract-content" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.053687 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="692fd2d3-4715-41af-9f10-7d11bebe6e86" containerName="registry-server" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.054638 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.079464 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-djvq2"] Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.142140 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fngwh\" (UniqueName: \"kubernetes.io/projected/459ff718-54ef-45f5-b32c-e144cf0c8f65-kube-api-access-fngwh\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.142240 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-catalog-content\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.142359 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-utilities\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.243948 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-utilities\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.244030 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fngwh\" (UniqueName: \"kubernetes.io/projected/459ff718-54ef-45f5-b32c-e144cf0c8f65-kube-api-access-fngwh\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.244098 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-catalog-content\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.244767 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-catalog-content\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.245570 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-utilities\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.277534 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fngwh\" (UniqueName: \"kubernetes.io/projected/459ff718-54ef-45f5-b32c-e144cf0c8f65-kube-api-access-fngwh\") pod \"community-operators-djvq2\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.378319 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:47:50 crc kubenswrapper[4946]: I1203 08:47:50.975070 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-djvq2"] Dec 03 08:47:51 crc kubenswrapper[4946]: I1203 08:47:51.532631 4946 generic.go:334] "Generic (PLEG): container finished" podID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerID="64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01" exitCode=0 Dec 03 08:47:51 crc kubenswrapper[4946]: I1203 08:47:51.532680 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-djvq2" event={"ID":"459ff718-54ef-45f5-b32c-e144cf0c8f65","Type":"ContainerDied","Data":"64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01"} Dec 03 08:47:51 crc kubenswrapper[4946]: I1203 08:47:51.534575 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-djvq2" event={"ID":"459ff718-54ef-45f5-b32c-e144cf0c8f65","Type":"ContainerStarted","Data":"0b68a2b681c2d739c0ee0952fee463deaac50847e5962ceecdec78890bc87345"} Dec 03 08:47:52 crc kubenswrapper[4946]: I1203 08:47:52.547526 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-djvq2" event={"ID":"459ff718-54ef-45f5-b32c-e144cf0c8f65","Type":"ContainerStarted","Data":"08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d"} Dec 03 08:47:53 crc kubenswrapper[4946]: I1203 08:47:53.039091 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:47:53 crc kubenswrapper[4946]: I1203 08:47:53.039475 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:47:53 crc kubenswrapper[4946]: I1203 08:47:53.581468 4946 generic.go:334] "Generic (PLEG): container finished" podID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerID="08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d" exitCode=0 Dec 03 08:47:53 crc kubenswrapper[4946]: I1203 08:47:53.581540 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-djvq2" event={"ID":"459ff718-54ef-45f5-b32c-e144cf0c8f65","Type":"ContainerDied","Data":"08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d"} Dec 03 08:47:53 crc kubenswrapper[4946]: I1203 08:47:53.581583 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-djvq2" event={"ID":"459ff718-54ef-45f5-b32c-e144cf0c8f65","Type":"ContainerStarted","Data":"6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01"} Dec 03 08:47:53 crc kubenswrapper[4946]: I1203 08:47:53.617010 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-djvq2" podStartSLOduration=2.158741438 podStartE2EDuration="3.616989038s" podCreationTimestamp="2025-12-03 08:47:50 +0000 UTC" firstStartedPulling="2025-12-03 08:47:51.534111917 +0000 UTC m=+7064.330802026" lastFinishedPulling="2025-12-03 08:47:52.992359497 +0000 UTC m=+7065.789049626" observedRunningTime="2025-12-03 08:47:53.614499513 +0000 UTC m=+7066.411189692" watchObservedRunningTime="2025-12-03 08:47:53.616989038 +0000 UTC m=+7066.413679147" Dec 03 08:48:00 crc kubenswrapper[4946]: I1203 08:48:00.379307 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:48:00 crc kubenswrapper[4946]: I1203 08:48:00.380089 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:48:00 crc kubenswrapper[4946]: I1203 08:48:00.469369 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:48:00 crc kubenswrapper[4946]: I1203 08:48:00.720410 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:48:00 crc kubenswrapper[4946]: I1203 08:48:00.792664 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-djvq2"] Dec 03 08:48:02 crc kubenswrapper[4946]: I1203 08:48:02.669801 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-djvq2" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerName="registry-server" containerID="cri-o://6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01" gracePeriod=2 Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.184988 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.255521 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-catalog-content\") pod \"459ff718-54ef-45f5-b32c-e144cf0c8f65\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.255672 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-utilities\") pod \"459ff718-54ef-45f5-b32c-e144cf0c8f65\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.255864 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fngwh\" (UniqueName: \"kubernetes.io/projected/459ff718-54ef-45f5-b32c-e144cf0c8f65-kube-api-access-fngwh\") pod \"459ff718-54ef-45f5-b32c-e144cf0c8f65\" (UID: \"459ff718-54ef-45f5-b32c-e144cf0c8f65\") " Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.256797 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-utilities" (OuterVolumeSpecName: "utilities") pod "459ff718-54ef-45f5-b32c-e144cf0c8f65" (UID: "459ff718-54ef-45f5-b32c-e144cf0c8f65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.262171 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/459ff718-54ef-45f5-b32c-e144cf0c8f65-kube-api-access-fngwh" (OuterVolumeSpecName: "kube-api-access-fngwh") pod "459ff718-54ef-45f5-b32c-e144cf0c8f65" (UID: "459ff718-54ef-45f5-b32c-e144cf0c8f65"). InnerVolumeSpecName "kube-api-access-fngwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.323175 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "459ff718-54ef-45f5-b32c-e144cf0c8f65" (UID: "459ff718-54ef-45f5-b32c-e144cf0c8f65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.357517 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fngwh\" (UniqueName: \"kubernetes.io/projected/459ff718-54ef-45f5-b32c-e144cf0c8f65-kube-api-access-fngwh\") on node \"crc\" DevicePath \"\"" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.357582 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.357596 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459ff718-54ef-45f5-b32c-e144cf0c8f65-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.678639 4946 generic.go:334] "Generic (PLEG): container finished" podID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerID="6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01" exitCode=0 Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.678695 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-djvq2" event={"ID":"459ff718-54ef-45f5-b32c-e144cf0c8f65","Type":"ContainerDied","Data":"6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01"} Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.678726 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-djvq2" event={"ID":"459ff718-54ef-45f5-b32c-e144cf0c8f65","Type":"ContainerDied","Data":"0b68a2b681c2d739c0ee0952fee463deaac50847e5962ceecdec78890bc87345"} Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.678759 4946 scope.go:117] "RemoveContainer" containerID="6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.678881 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-djvq2" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.699593 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-djvq2"] Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.707118 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-djvq2"] Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.717216 4946 scope.go:117] "RemoveContainer" containerID="08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.742515 4946 scope.go:117] "RemoveContainer" containerID="64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.789299 4946 scope.go:117] "RemoveContainer" containerID="6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01" Dec 03 08:48:03 crc kubenswrapper[4946]: E1203 08:48:03.789963 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01\": container with ID starting with 6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01 not found: ID does not exist" containerID="6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.790067 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01"} err="failed to get container status \"6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01\": rpc error: code = NotFound desc = could not find container \"6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01\": container with ID starting with 6354c2a547686695ab3ccc8c65f0f46e6a052b1ab1cddc856da54799b40f2f01 not found: ID does not exist" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.790136 4946 scope.go:117] "RemoveContainer" containerID="08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d" Dec 03 08:48:03 crc kubenswrapper[4946]: E1203 08:48:03.790755 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d\": container with ID starting with 08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d not found: ID does not exist" containerID="08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.790790 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d"} err="failed to get container status \"08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d\": rpc error: code = NotFound desc = could not find container \"08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d\": container with ID starting with 08369e3fbb5148618b7cd0bf1f1d96173a8957e6bff2b01e59c398b0c1e1cd9d not found: ID does not exist" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.790815 4946 scope.go:117] "RemoveContainer" containerID="64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01" Dec 03 08:48:03 crc kubenswrapper[4946]: E1203 08:48:03.791147 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01\": container with ID starting with 64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01 not found: ID does not exist" containerID="64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01" Dec 03 08:48:03 crc kubenswrapper[4946]: I1203 08:48:03.791230 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01"} err="failed to get container status \"64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01\": rpc error: code = NotFound desc = could not find container \"64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01\": container with ID starting with 64d8febbc530be06efadf74a500cf090ff660dafb5d40a150338f2d9413d7f01 not found: ID does not exist" Dec 03 08:48:05 crc kubenswrapper[4946]: I1203 08:48:05.601925 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" path="/var/lib/kubelet/pods/459ff718-54ef-45f5-b32c-e144cf0c8f65/volumes" Dec 03 08:48:23 crc kubenswrapper[4946]: I1203 08:48:23.040219 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:48:23 crc kubenswrapper[4946]: I1203 08:48:23.041208 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:48:53 crc kubenswrapper[4946]: I1203 08:48:53.039562 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:48:53 crc kubenswrapper[4946]: I1203 08:48:53.040264 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:48:53 crc kubenswrapper[4946]: I1203 08:48:53.040339 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:48:53 crc kubenswrapper[4946]: I1203 08:48:53.041426 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:48:53 crc kubenswrapper[4946]: I1203 08:48:53.041555 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" gracePeriod=600 Dec 03 08:48:53 crc kubenswrapper[4946]: I1203 08:48:53.173139 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" exitCode=0 Dec 03 08:48:53 crc kubenswrapper[4946]: I1203 08:48:53.173190 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad"} Dec 03 08:48:53 crc kubenswrapper[4946]: I1203 08:48:53.173239 4946 scope.go:117] "RemoveContainer" containerID="878aee7b8ce5cbe55128d3e3c0dca6f3787d24f759d3a27c81faad7b98362df5" Dec 03 08:48:53 crc kubenswrapper[4946]: E1203 08:48:53.173664 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:48:54 crc kubenswrapper[4946]: I1203 08:48:54.186531 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:48:54 crc kubenswrapper[4946]: E1203 08:48:54.187157 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:49:05 crc kubenswrapper[4946]: I1203 08:49:05.593574 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:49:05 crc kubenswrapper[4946]: E1203 08:49:05.594589 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.534542 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g7kjj"] Dec 03 08:49:08 crc kubenswrapper[4946]: E1203 08:49:08.535223 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerName="extract-utilities" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.535238 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerName="extract-utilities" Dec 03 08:49:08 crc kubenswrapper[4946]: E1203 08:49:08.535261 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerName="registry-server" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.535271 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerName="registry-server" Dec 03 08:49:08 crc kubenswrapper[4946]: E1203 08:49:08.535296 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerName="extract-content" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.535307 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerName="extract-content" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.535566 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="459ff718-54ef-45f5-b32c-e144cf0c8f65" containerName="registry-server" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.540032 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.546166 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g7kjj"] Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.725179 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km4bq\" (UniqueName: \"kubernetes.io/projected/41c7c356-300a-4f80-8428-29766be9b61d-kube-api-access-km4bq\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.725472 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-utilities\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.725583 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-catalog-content\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.827007 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-utilities\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.827071 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-catalog-content\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.827157 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km4bq\" (UniqueName: \"kubernetes.io/projected/41c7c356-300a-4f80-8428-29766be9b61d-kube-api-access-km4bq\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.827881 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-utilities\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.827991 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-catalog-content\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:08 crc kubenswrapper[4946]: I1203 08:49:08.865924 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km4bq\" (UniqueName: \"kubernetes.io/projected/41c7c356-300a-4f80-8428-29766be9b61d-kube-api-access-km4bq\") pod \"redhat-operators-g7kjj\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:09 crc kubenswrapper[4946]: I1203 08:49:09.160435 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:09 crc kubenswrapper[4946]: I1203 08:49:09.407728 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g7kjj"] Dec 03 08:49:10 crc kubenswrapper[4946]: I1203 08:49:10.325489 4946 generic.go:334] "Generic (PLEG): container finished" podID="41c7c356-300a-4f80-8428-29766be9b61d" containerID="14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514" exitCode=0 Dec 03 08:49:10 crc kubenswrapper[4946]: I1203 08:49:10.325538 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7kjj" event={"ID":"41c7c356-300a-4f80-8428-29766be9b61d","Type":"ContainerDied","Data":"14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514"} Dec 03 08:49:10 crc kubenswrapper[4946]: I1203 08:49:10.325572 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7kjj" event={"ID":"41c7c356-300a-4f80-8428-29766be9b61d","Type":"ContainerStarted","Data":"237cef07d9ab2ecf0c1d1eb5e9270058af76182e34bb55c07da360ccfad0a53e"} Dec 03 08:49:12 crc kubenswrapper[4946]: I1203 08:49:12.343497 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7kjj" event={"ID":"41c7c356-300a-4f80-8428-29766be9b61d","Type":"ContainerStarted","Data":"805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8"} Dec 03 08:49:13 crc kubenswrapper[4946]: I1203 08:49:13.356034 4946 generic.go:334] "Generic (PLEG): container finished" podID="41c7c356-300a-4f80-8428-29766be9b61d" containerID="805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8" exitCode=0 Dec 03 08:49:13 crc kubenswrapper[4946]: I1203 08:49:13.356131 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7kjj" event={"ID":"41c7c356-300a-4f80-8428-29766be9b61d","Type":"ContainerDied","Data":"805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8"} Dec 03 08:49:14 crc kubenswrapper[4946]: I1203 08:49:14.367216 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7kjj" event={"ID":"41c7c356-300a-4f80-8428-29766be9b61d","Type":"ContainerStarted","Data":"b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4"} Dec 03 08:49:14 crc kubenswrapper[4946]: I1203 08:49:14.400835 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g7kjj" podStartSLOduration=2.937717683 podStartE2EDuration="6.400775937s" podCreationTimestamp="2025-12-03 08:49:08 +0000 UTC" firstStartedPulling="2025-12-03 08:49:10.326936757 +0000 UTC m=+7143.123626866" lastFinishedPulling="2025-12-03 08:49:13.789995011 +0000 UTC m=+7146.586685120" observedRunningTime="2025-12-03 08:49:14.391112093 +0000 UTC m=+7147.187802212" watchObservedRunningTime="2025-12-03 08:49:14.400775937 +0000 UTC m=+7147.197466086" Dec 03 08:49:16 crc kubenswrapper[4946]: I1203 08:49:16.592960 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:49:16 crc kubenswrapper[4946]: E1203 08:49:16.593897 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:49:19 crc kubenswrapper[4946]: I1203 08:49:19.161480 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:19 crc kubenswrapper[4946]: I1203 08:49:19.161888 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:20 crc kubenswrapper[4946]: I1203 08:49:20.228174 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g7kjj" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="registry-server" probeResult="failure" output=< Dec 03 08:49:20 crc kubenswrapper[4946]: timeout: failed to connect service ":50051" within 1s Dec 03 08:49:20 crc kubenswrapper[4946]: > Dec 03 08:49:28 crc kubenswrapper[4946]: I1203 08:49:28.593358 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:49:28 crc kubenswrapper[4946]: E1203 08:49:28.594456 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:49:29 crc kubenswrapper[4946]: I1203 08:49:29.285600 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:29 crc kubenswrapper[4946]: I1203 08:49:29.331336 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:29 crc kubenswrapper[4946]: I1203 08:49:29.519846 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g7kjj"] Dec 03 08:49:30 crc kubenswrapper[4946]: I1203 08:49:30.506946 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g7kjj" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="registry-server" containerID="cri-o://b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4" gracePeriod=2 Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.021679 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.206245 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km4bq\" (UniqueName: \"kubernetes.io/projected/41c7c356-300a-4f80-8428-29766be9b61d-kube-api-access-km4bq\") pod \"41c7c356-300a-4f80-8428-29766be9b61d\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.206308 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-catalog-content\") pod \"41c7c356-300a-4f80-8428-29766be9b61d\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.206346 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-utilities\") pod \"41c7c356-300a-4f80-8428-29766be9b61d\" (UID: \"41c7c356-300a-4f80-8428-29766be9b61d\") " Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.208027 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-utilities" (OuterVolumeSpecName: "utilities") pod "41c7c356-300a-4f80-8428-29766be9b61d" (UID: "41c7c356-300a-4f80-8428-29766be9b61d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.218477 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41c7c356-300a-4f80-8428-29766be9b61d-kube-api-access-km4bq" (OuterVolumeSpecName: "kube-api-access-km4bq") pod "41c7c356-300a-4f80-8428-29766be9b61d" (UID: "41c7c356-300a-4f80-8428-29766be9b61d"). InnerVolumeSpecName "kube-api-access-km4bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.307859 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km4bq\" (UniqueName: \"kubernetes.io/projected/41c7c356-300a-4f80-8428-29766be9b61d-kube-api-access-km4bq\") on node \"crc\" DevicePath \"\"" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.307896 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.356836 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41c7c356-300a-4f80-8428-29766be9b61d" (UID: "41c7c356-300a-4f80-8428-29766be9b61d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.409456 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c356-300a-4f80-8428-29766be9b61d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.516605 4946 generic.go:334] "Generic (PLEG): container finished" podID="41c7c356-300a-4f80-8428-29766be9b61d" containerID="b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4" exitCode=0 Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.516685 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7kjj" event={"ID":"41c7c356-300a-4f80-8428-29766be9b61d","Type":"ContainerDied","Data":"b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4"} Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.516720 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7kjj" event={"ID":"41c7c356-300a-4f80-8428-29766be9b61d","Type":"ContainerDied","Data":"237cef07d9ab2ecf0c1d1eb5e9270058af76182e34bb55c07da360ccfad0a53e"} Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.516755 4946 scope.go:117] "RemoveContainer" containerID="b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.516799 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7kjj" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.541405 4946 scope.go:117] "RemoveContainer" containerID="805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.561883 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g7kjj"] Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.568306 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g7kjj"] Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.575260 4946 scope.go:117] "RemoveContainer" containerID="14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.602398 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41c7c356-300a-4f80-8428-29766be9b61d" path="/var/lib/kubelet/pods/41c7c356-300a-4f80-8428-29766be9b61d/volumes" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.605465 4946 scope.go:117] "RemoveContainer" containerID="b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4" Dec 03 08:49:31 crc kubenswrapper[4946]: E1203 08:49:31.606009 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4\": container with ID starting with b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4 not found: ID does not exist" containerID="b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.606067 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4"} err="failed to get container status \"b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4\": rpc error: code = NotFound desc = could not find container \"b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4\": container with ID starting with b6f5b76e7ad5c0b4124022670f568c22f4b4b232b1ca7283efd9685a7627bfe4 not found: ID does not exist" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.606103 4946 scope.go:117] "RemoveContainer" containerID="805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8" Dec 03 08:49:31 crc kubenswrapper[4946]: E1203 08:49:31.606496 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8\": container with ID starting with 805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8 not found: ID does not exist" containerID="805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.606532 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8"} err="failed to get container status \"805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8\": rpc error: code = NotFound desc = could not find container \"805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8\": container with ID starting with 805327ac73f8a90fae694ac3fc57eff98213889f295344f6d59d20388d8805a8 not found: ID does not exist" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.606556 4946 scope.go:117] "RemoveContainer" containerID="14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514" Dec 03 08:49:31 crc kubenswrapper[4946]: E1203 08:49:31.607206 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514\": container with ID starting with 14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514 not found: ID does not exist" containerID="14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514" Dec 03 08:49:31 crc kubenswrapper[4946]: I1203 08:49:31.607246 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514"} err="failed to get container status \"14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514\": rpc error: code = NotFound desc = could not find container \"14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514\": container with ID starting with 14ca1b4e9e9ff81ef16ab5df4d40c04fc0f5cbf387e8a6c50db243e365fcc514 not found: ID does not exist" Dec 03 08:49:40 crc kubenswrapper[4946]: I1203 08:49:40.593506 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:49:40 crc kubenswrapper[4946]: E1203 08:49:40.594440 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:49:51 crc kubenswrapper[4946]: I1203 08:49:51.593098 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:49:51 crc kubenswrapper[4946]: E1203 08:49:51.594055 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:50:04 crc kubenswrapper[4946]: I1203 08:50:04.593124 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:50:04 crc kubenswrapper[4946]: E1203 08:50:04.594382 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:50:18 crc kubenswrapper[4946]: I1203 08:50:18.593643 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:50:18 crc kubenswrapper[4946]: E1203 08:50:18.594508 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:50:32 crc kubenswrapper[4946]: I1203 08:50:32.592158 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:50:32 crc kubenswrapper[4946]: E1203 08:50:32.592955 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:50:44 crc kubenswrapper[4946]: I1203 08:50:44.592527 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:50:44 crc kubenswrapper[4946]: E1203 08:50:44.593320 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:50:59 crc kubenswrapper[4946]: I1203 08:50:59.596143 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:50:59 crc kubenswrapper[4946]: E1203 08:50:59.597134 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:51:14 crc kubenswrapper[4946]: I1203 08:51:14.592692 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:51:14 crc kubenswrapper[4946]: E1203 08:51:14.593881 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:51:25 crc kubenswrapper[4946]: I1203 08:51:25.593817 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:51:25 crc kubenswrapper[4946]: E1203 08:51:25.595212 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:51:40 crc kubenswrapper[4946]: I1203 08:51:40.593781 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:51:40 crc kubenswrapper[4946]: E1203 08:51:40.595017 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:51:54 crc kubenswrapper[4946]: I1203 08:51:54.593937 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:51:54 crc kubenswrapper[4946]: E1203 08:51:54.595336 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:52:06 crc kubenswrapper[4946]: I1203 08:52:06.593282 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:52:06 crc kubenswrapper[4946]: E1203 08:52:06.594117 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:52:17 crc kubenswrapper[4946]: I1203 08:52:17.598277 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:52:17 crc kubenswrapper[4946]: E1203 08:52:17.599500 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:52:31 crc kubenswrapper[4946]: I1203 08:52:31.593434 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:52:31 crc kubenswrapper[4946]: E1203 08:52:31.594604 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:52:44 crc kubenswrapper[4946]: I1203 08:52:44.592694 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:52:44 crc kubenswrapper[4946]: E1203 08:52:44.593800 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:52:55 crc kubenswrapper[4946]: I1203 08:52:55.593447 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:52:55 crc kubenswrapper[4946]: E1203 08:52:55.594827 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:53:10 crc kubenswrapper[4946]: I1203 08:53:10.593511 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:53:10 crc kubenswrapper[4946]: E1203 08:53:10.594883 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:53:23 crc kubenswrapper[4946]: I1203 08:53:23.592843 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:53:23 crc kubenswrapper[4946]: E1203 08:53:23.593817 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:53:36 crc kubenswrapper[4946]: I1203 08:53:36.593143 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:53:36 crc kubenswrapper[4946]: E1203 08:53:36.594358 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:53:48 crc kubenswrapper[4946]: I1203 08:53:48.593367 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:53:48 crc kubenswrapper[4946]: E1203 08:53:48.594416 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 08:53:59 crc kubenswrapper[4946]: I1203 08:53:59.593102 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:54:00 crc kubenswrapper[4946]: I1203 08:54:00.039080 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"6ec43767614acc0b2c9ae69b6756fbf948a7fd22b3606e6c66341c4c8adda2fe"} Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.884526 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l75gx"] Dec 03 08:56:07 crc kubenswrapper[4946]: E1203 08:56:07.885457 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="extract-utilities" Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.885472 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="extract-utilities" Dec 03 08:56:07 crc kubenswrapper[4946]: E1203 08:56:07.885493 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="extract-content" Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.885501 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="extract-content" Dec 03 08:56:07 crc kubenswrapper[4946]: E1203 08:56:07.885524 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="registry-server" Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.885531 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="registry-server" Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.885707 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="41c7c356-300a-4f80-8428-29766be9b61d" containerName="registry-server" Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.887010 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.907306 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l75gx"] Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.965223 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-catalog-content\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.965347 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shzhm\" (UniqueName: \"kubernetes.io/projected/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-kube-api-access-shzhm\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:07 crc kubenswrapper[4946]: I1203 08:56:07.965377 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-utilities\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:08 crc kubenswrapper[4946]: I1203 08:56:08.067837 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-catalog-content\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:08 crc kubenswrapper[4946]: I1203 08:56:08.068027 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shzhm\" (UniqueName: \"kubernetes.io/projected/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-kube-api-access-shzhm\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:08 crc kubenswrapper[4946]: I1203 08:56:08.068096 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-utilities\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:08 crc kubenswrapper[4946]: I1203 08:56:08.068388 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-catalog-content\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:08 crc kubenswrapper[4946]: I1203 08:56:08.068627 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-utilities\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:08 crc kubenswrapper[4946]: I1203 08:56:08.091450 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shzhm\" (UniqueName: \"kubernetes.io/projected/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-kube-api-access-shzhm\") pod \"certified-operators-l75gx\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:08 crc kubenswrapper[4946]: I1203 08:56:08.209019 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:08 crc kubenswrapper[4946]: I1203 08:56:08.724848 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l75gx"] Dec 03 08:56:09 crc kubenswrapper[4946]: I1203 08:56:09.191380 4946 generic.go:334] "Generic (PLEG): container finished" podID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerID="175c65a4c6f4fc0ed8a1795df0879149ca19f82338c58af3dfafa34c3fafe61c" exitCode=0 Dec 03 08:56:09 crc kubenswrapper[4946]: I1203 08:56:09.191530 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l75gx" event={"ID":"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480","Type":"ContainerDied","Data":"175c65a4c6f4fc0ed8a1795df0879149ca19f82338c58af3dfafa34c3fafe61c"} Dec 03 08:56:09 crc kubenswrapper[4946]: I1203 08:56:09.192052 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l75gx" event={"ID":"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480","Type":"ContainerStarted","Data":"72e326912f229145bc1bf6df170cb6bea5643dccec5b71d50b8b96fadc117733"} Dec 03 08:56:09 crc kubenswrapper[4946]: I1203 08:56:09.196685 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:56:11 crc kubenswrapper[4946]: I1203 08:56:11.212661 4946 generic.go:334] "Generic (PLEG): container finished" podID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerID="1cd2d25457bff833f29692553c95400c5ce7fa215540c32a200fa8d6bc1178eb" exitCode=0 Dec 03 08:56:11 crc kubenswrapper[4946]: I1203 08:56:11.212732 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l75gx" event={"ID":"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480","Type":"ContainerDied","Data":"1cd2d25457bff833f29692553c95400c5ce7fa215540c32a200fa8d6bc1178eb"} Dec 03 08:56:13 crc kubenswrapper[4946]: I1203 08:56:13.230861 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l75gx" event={"ID":"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480","Type":"ContainerStarted","Data":"d494b600b2d64f5a5bfbe2a8a152a384329e83ed68d4d55dd6f6bf949061c751"} Dec 03 08:56:13 crc kubenswrapper[4946]: I1203 08:56:13.254688 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l75gx" podStartSLOduration=3.3472887350000002 podStartE2EDuration="6.254655483s" podCreationTimestamp="2025-12-03 08:56:07 +0000 UTC" firstStartedPulling="2025-12-03 08:56:09.195888782 +0000 UTC m=+7561.992578931" lastFinishedPulling="2025-12-03 08:56:12.10325553 +0000 UTC m=+7564.899945679" observedRunningTime="2025-12-03 08:56:13.249826807 +0000 UTC m=+7566.046516926" watchObservedRunningTime="2025-12-03 08:56:13.254655483 +0000 UTC m=+7566.051345592" Dec 03 08:56:18 crc kubenswrapper[4946]: I1203 08:56:18.209859 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:18 crc kubenswrapper[4946]: I1203 08:56:18.210539 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:18 crc kubenswrapper[4946]: I1203 08:56:18.279960 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:18 crc kubenswrapper[4946]: I1203 08:56:18.356399 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:18 crc kubenswrapper[4946]: I1203 08:56:18.521616 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l75gx"] Dec 03 08:56:20 crc kubenswrapper[4946]: I1203 08:56:20.293095 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l75gx" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerName="registry-server" containerID="cri-o://d494b600b2d64f5a5bfbe2a8a152a384329e83ed68d4d55dd6f6bf949061c751" gracePeriod=2 Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.318505 4946 generic.go:334] "Generic (PLEG): container finished" podID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerID="d494b600b2d64f5a5bfbe2a8a152a384329e83ed68d4d55dd6f6bf949061c751" exitCode=0 Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.318969 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l75gx" event={"ID":"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480","Type":"ContainerDied","Data":"d494b600b2d64f5a5bfbe2a8a152a384329e83ed68d4d55dd6f6bf949061c751"} Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.466851 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.591883 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-catalog-content\") pod \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.592044 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shzhm\" (UniqueName: \"kubernetes.io/projected/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-kube-api-access-shzhm\") pod \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.592076 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-utilities\") pod \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\" (UID: \"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480\") " Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.593139 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-utilities" (OuterVolumeSpecName: "utilities") pod "e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" (UID: "e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.602003 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-kube-api-access-shzhm" (OuterVolumeSpecName: "kube-api-access-shzhm") pod "e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" (UID: "e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480"). InnerVolumeSpecName "kube-api-access-shzhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.641509 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" (UID: "e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.693309 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shzhm\" (UniqueName: \"kubernetes.io/projected/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-kube-api-access-shzhm\") on node \"crc\" DevicePath \"\"" Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.693340 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:56:21 crc kubenswrapper[4946]: I1203 08:56:21.693352 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:56:22 crc kubenswrapper[4946]: I1203 08:56:22.331568 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l75gx" event={"ID":"e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480","Type":"ContainerDied","Data":"72e326912f229145bc1bf6df170cb6bea5643dccec5b71d50b8b96fadc117733"} Dec 03 08:56:22 crc kubenswrapper[4946]: I1203 08:56:22.331636 4946 scope.go:117] "RemoveContainer" containerID="d494b600b2d64f5a5bfbe2a8a152a384329e83ed68d4d55dd6f6bf949061c751" Dec 03 08:56:22 crc kubenswrapper[4946]: I1203 08:56:22.331729 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l75gx" Dec 03 08:56:22 crc kubenswrapper[4946]: I1203 08:56:22.353963 4946 scope.go:117] "RemoveContainer" containerID="1cd2d25457bff833f29692553c95400c5ce7fa215540c32a200fa8d6bc1178eb" Dec 03 08:56:22 crc kubenswrapper[4946]: I1203 08:56:22.367999 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l75gx"] Dec 03 08:56:22 crc kubenswrapper[4946]: I1203 08:56:22.386658 4946 scope.go:117] "RemoveContainer" containerID="175c65a4c6f4fc0ed8a1795df0879149ca19f82338c58af3dfafa34c3fafe61c" Dec 03 08:56:22 crc kubenswrapper[4946]: I1203 08:56:22.388958 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l75gx"] Dec 03 08:56:23 crc kubenswrapper[4946]: I1203 08:56:23.039131 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:56:23 crc kubenswrapper[4946]: I1203 08:56:23.039503 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:56:23 crc kubenswrapper[4946]: I1203 08:56:23.601222 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" path="/var/lib/kubelet/pods/e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480/volumes" Dec 03 08:56:53 crc kubenswrapper[4946]: I1203 08:56:53.039968 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:56:53 crc kubenswrapper[4946]: I1203 08:56:53.040632 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.234958 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2kdsb"] Dec 03 08:56:55 crc kubenswrapper[4946]: E1203 08:56:55.235848 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerName="registry-server" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.235872 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerName="registry-server" Dec 03 08:56:55 crc kubenswrapper[4946]: E1203 08:56:55.235900 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerName="extract-content" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.235911 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerName="extract-content" Dec 03 08:56:55 crc kubenswrapper[4946]: E1203 08:56:55.235960 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerName="extract-utilities" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.235971 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerName="extract-utilities" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.236183 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b4d5f5-6ae7-4c81-bfdb-26d18d98d480" containerName="registry-server" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.241193 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.259888 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kdsb"] Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.394080 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-utilities\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.394160 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6q78\" (UniqueName: \"kubernetes.io/projected/797faf4a-9249-4c8a-a22f-41255fe8ebce-kube-api-access-t6q78\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.394253 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-catalog-content\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.496724 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-utilities\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.496817 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6q78\" (UniqueName: \"kubernetes.io/projected/797faf4a-9249-4c8a-a22f-41255fe8ebce-kube-api-access-t6q78\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.496855 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-catalog-content\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.497317 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-utilities\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.497335 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-catalog-content\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.519786 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6q78\" (UniqueName: \"kubernetes.io/projected/797faf4a-9249-4c8a-a22f-41255fe8ebce-kube-api-access-t6q78\") pod \"redhat-marketplace-2kdsb\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:55 crc kubenswrapper[4946]: I1203 08:56:55.566638 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:56:56 crc kubenswrapper[4946]: I1203 08:56:56.055987 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kdsb"] Dec 03 08:56:56 crc kubenswrapper[4946]: I1203 08:56:56.634608 4946 generic.go:334] "Generic (PLEG): container finished" podID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerID="a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588" exitCode=0 Dec 03 08:56:56 crc kubenswrapper[4946]: I1203 08:56:56.634675 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kdsb" event={"ID":"797faf4a-9249-4c8a-a22f-41255fe8ebce","Type":"ContainerDied","Data":"a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588"} Dec 03 08:56:56 crc kubenswrapper[4946]: I1203 08:56:56.634719 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kdsb" event={"ID":"797faf4a-9249-4c8a-a22f-41255fe8ebce","Type":"ContainerStarted","Data":"21a70a456f577b8589f970e2252d09b4dcf3ff2375fb25d04acbb1bbf1870825"} Dec 03 08:56:57 crc kubenswrapper[4946]: I1203 08:56:57.645405 4946 generic.go:334] "Generic (PLEG): container finished" podID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerID="fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7" exitCode=0 Dec 03 08:56:57 crc kubenswrapper[4946]: I1203 08:56:57.645550 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kdsb" event={"ID":"797faf4a-9249-4c8a-a22f-41255fe8ebce","Type":"ContainerDied","Data":"fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7"} Dec 03 08:56:58 crc kubenswrapper[4946]: I1203 08:56:58.661497 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kdsb" event={"ID":"797faf4a-9249-4c8a-a22f-41255fe8ebce","Type":"ContainerStarted","Data":"6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836"} Dec 03 08:56:58 crc kubenswrapper[4946]: I1203 08:56:58.688558 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2kdsb" podStartSLOduration=2.23438623 podStartE2EDuration="3.688533382s" podCreationTimestamp="2025-12-03 08:56:55 +0000 UTC" firstStartedPulling="2025-12-03 08:56:56.636991567 +0000 UTC m=+7609.433681676" lastFinishedPulling="2025-12-03 08:56:58.091138679 +0000 UTC m=+7610.887828828" observedRunningTime="2025-12-03 08:56:58.686277663 +0000 UTC m=+7611.482967852" watchObservedRunningTime="2025-12-03 08:56:58.688533382 +0000 UTC m=+7611.485223501" Dec 03 08:57:05 crc kubenswrapper[4946]: I1203 08:57:05.566843 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:57:05 crc kubenswrapper[4946]: I1203 08:57:05.567247 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:57:05 crc kubenswrapper[4946]: I1203 08:57:05.610866 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:57:05 crc kubenswrapper[4946]: I1203 08:57:05.775411 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:57:05 crc kubenswrapper[4946]: I1203 08:57:05.845855 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kdsb"] Dec 03 08:57:07 crc kubenswrapper[4946]: I1203 08:57:07.746801 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2kdsb" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerName="registry-server" containerID="cri-o://6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836" gracePeriod=2 Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.737127 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.759389 4946 generic.go:334] "Generic (PLEG): container finished" podID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerID="6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836" exitCode=0 Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.759451 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kdsb" event={"ID":"797faf4a-9249-4c8a-a22f-41255fe8ebce","Type":"ContainerDied","Data":"6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836"} Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.759489 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kdsb" event={"ID":"797faf4a-9249-4c8a-a22f-41255fe8ebce","Type":"ContainerDied","Data":"21a70a456f577b8589f970e2252d09b4dcf3ff2375fb25d04acbb1bbf1870825"} Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.759508 4946 scope.go:117] "RemoveContainer" containerID="6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.759450 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kdsb" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.785219 4946 scope.go:117] "RemoveContainer" containerID="fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.805446 4946 scope.go:117] "RemoveContainer" containerID="a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.830015 4946 scope.go:117] "RemoveContainer" containerID="6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836" Dec 03 08:57:08 crc kubenswrapper[4946]: E1203 08:57:08.830573 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836\": container with ID starting with 6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836 not found: ID does not exist" containerID="6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.830675 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836"} err="failed to get container status \"6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836\": rpc error: code = NotFound desc = could not find container \"6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836\": container with ID starting with 6839333cb9d24df854875b090511731ab8a2a85f06fc3128cc0a5317667a4836 not found: ID does not exist" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.830731 4946 scope.go:117] "RemoveContainer" containerID="fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7" Dec 03 08:57:08 crc kubenswrapper[4946]: E1203 08:57:08.831110 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7\": container with ID starting with fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7 not found: ID does not exist" containerID="fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.831140 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7"} err="failed to get container status \"fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7\": rpc error: code = NotFound desc = could not find container \"fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7\": container with ID starting with fddb3ae8144a1ea9e132a4faf318870e22a7a38f8e50502a91398f9990f983a7 not found: ID does not exist" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.831158 4946 scope.go:117] "RemoveContainer" containerID="a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588" Dec 03 08:57:08 crc kubenswrapper[4946]: E1203 08:57:08.831515 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588\": container with ID starting with a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588 not found: ID does not exist" containerID="a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.831543 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588"} err="failed to get container status \"a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588\": rpc error: code = NotFound desc = could not find container \"a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588\": container with ID starting with a8816876d30ce83384b7e383441220da992afe2a32350b72f042d31cd7dff588 not found: ID does not exist" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.905731 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6q78\" (UniqueName: \"kubernetes.io/projected/797faf4a-9249-4c8a-a22f-41255fe8ebce-kube-api-access-t6q78\") pod \"797faf4a-9249-4c8a-a22f-41255fe8ebce\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.905883 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-catalog-content\") pod \"797faf4a-9249-4c8a-a22f-41255fe8ebce\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.905987 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-utilities\") pod \"797faf4a-9249-4c8a-a22f-41255fe8ebce\" (UID: \"797faf4a-9249-4c8a-a22f-41255fe8ebce\") " Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.907855 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-utilities" (OuterVolumeSpecName: "utilities") pod "797faf4a-9249-4c8a-a22f-41255fe8ebce" (UID: "797faf4a-9249-4c8a-a22f-41255fe8ebce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.914777 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797faf4a-9249-4c8a-a22f-41255fe8ebce-kube-api-access-t6q78" (OuterVolumeSpecName: "kube-api-access-t6q78") pod "797faf4a-9249-4c8a-a22f-41255fe8ebce" (UID: "797faf4a-9249-4c8a-a22f-41255fe8ebce"). InnerVolumeSpecName "kube-api-access-t6q78". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:57:08 crc kubenswrapper[4946]: I1203 08:57:08.936326 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "797faf4a-9249-4c8a-a22f-41255fe8ebce" (UID: "797faf4a-9249-4c8a-a22f-41255fe8ebce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:57:09 crc kubenswrapper[4946]: I1203 08:57:09.007812 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6q78\" (UniqueName: \"kubernetes.io/projected/797faf4a-9249-4c8a-a22f-41255fe8ebce-kube-api-access-t6q78\") on node \"crc\" DevicePath \"\"" Dec 03 08:57:09 crc kubenswrapper[4946]: I1203 08:57:09.007847 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:57:09 crc kubenswrapper[4946]: I1203 08:57:09.007857 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/797faf4a-9249-4c8a-a22f-41255fe8ebce-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:57:09 crc kubenswrapper[4946]: I1203 08:57:09.118664 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kdsb"] Dec 03 08:57:09 crc kubenswrapper[4946]: I1203 08:57:09.125831 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kdsb"] Dec 03 08:57:09 crc kubenswrapper[4946]: I1203 08:57:09.611488 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" path="/var/lib/kubelet/pods/797faf4a-9249-4c8a-a22f-41255fe8ebce/volumes" Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.039887 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.040638 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.040703 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.041731 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ec43767614acc0b2c9ae69b6756fbf948a7fd22b3606e6c66341c4c8adda2fe"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.041869 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://6ec43767614acc0b2c9ae69b6756fbf948a7fd22b3606e6c66341c4c8adda2fe" gracePeriod=600 Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.895227 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="6ec43767614acc0b2c9ae69b6756fbf948a7fd22b3606e6c66341c4c8adda2fe" exitCode=0 Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.895301 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"6ec43767614acc0b2c9ae69b6756fbf948a7fd22b3606e6c66341c4c8adda2fe"} Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.895703 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6"} Dec 03 08:57:23 crc kubenswrapper[4946]: I1203 08:57:23.895847 4946 scope.go:117] "RemoveContainer" containerID="1523e238a55335d5d60dfeba85a7d0e92b2df772fc1cb6c684e636ef388a19ad" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.000007 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2xd8h"] Dec 03 08:58:44 crc kubenswrapper[4946]: E1203 08:58:44.001310 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerName="registry-server" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.001529 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerName="registry-server" Dec 03 08:58:44 crc kubenswrapper[4946]: E1203 08:58:44.001584 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerName="extract-utilities" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.001604 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerName="extract-utilities" Dec 03 08:58:44 crc kubenswrapper[4946]: E1203 08:58:44.001646 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerName="extract-content" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.001666 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerName="extract-content" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.002054 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="797faf4a-9249-4c8a-a22f-41255fe8ebce" containerName="registry-server" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.004620 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.011296 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2xd8h"] Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.167609 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-catalog-content\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.167921 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss9t5\" (UniqueName: \"kubernetes.io/projected/780e9ab7-484c-489d-8594-14f42bd8d8f8-kube-api-access-ss9t5\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.168106 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-utilities\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.269700 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-catalog-content\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.269770 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss9t5\" (UniqueName: \"kubernetes.io/projected/780e9ab7-484c-489d-8594-14f42bd8d8f8-kube-api-access-ss9t5\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.269838 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-utilities\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.270522 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-utilities\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.270532 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-catalog-content\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.297670 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss9t5\" (UniqueName: \"kubernetes.io/projected/780e9ab7-484c-489d-8594-14f42bd8d8f8-kube-api-access-ss9t5\") pod \"community-operators-2xd8h\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.326246 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:44 crc kubenswrapper[4946]: I1203 08:58:44.836704 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2xd8h"] Dec 03 08:58:44 crc kubenswrapper[4946]: W1203 08:58:44.854693 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod780e9ab7_484c_489d_8594_14f42bd8d8f8.slice/crio-2cd6a8001027157e5f91946cc10ab16430fa054411e3a2ca6ff9334ca7aa4914 WatchSource:0}: Error finding container 2cd6a8001027157e5f91946cc10ab16430fa054411e3a2ca6ff9334ca7aa4914: Status 404 returned error can't find the container with id 2cd6a8001027157e5f91946cc10ab16430fa054411e3a2ca6ff9334ca7aa4914 Dec 03 08:58:45 crc kubenswrapper[4946]: I1203 08:58:45.720538 4946 generic.go:334] "Generic (PLEG): container finished" podID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerID="2e1bfd444f70df7cf550e6e0c8074f3f02a8d60d17631b523cfc5834a1e63373" exitCode=0 Dec 03 08:58:45 crc kubenswrapper[4946]: I1203 08:58:45.720632 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xd8h" event={"ID":"780e9ab7-484c-489d-8594-14f42bd8d8f8","Type":"ContainerDied","Data":"2e1bfd444f70df7cf550e6e0c8074f3f02a8d60d17631b523cfc5834a1e63373"} Dec 03 08:58:45 crc kubenswrapper[4946]: I1203 08:58:45.720936 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xd8h" event={"ID":"780e9ab7-484c-489d-8594-14f42bd8d8f8","Type":"ContainerStarted","Data":"2cd6a8001027157e5f91946cc10ab16430fa054411e3a2ca6ff9334ca7aa4914"} Dec 03 08:58:46 crc kubenswrapper[4946]: I1203 08:58:46.736531 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xd8h" event={"ID":"780e9ab7-484c-489d-8594-14f42bd8d8f8","Type":"ContainerStarted","Data":"0273836849c25c7e2bc31121d9b71d28fd05a7a9734b1afecbd39dfaddad2643"} Dec 03 08:58:47 crc kubenswrapper[4946]: I1203 08:58:47.744683 4946 generic.go:334] "Generic (PLEG): container finished" podID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerID="0273836849c25c7e2bc31121d9b71d28fd05a7a9734b1afecbd39dfaddad2643" exitCode=0 Dec 03 08:58:47 crc kubenswrapper[4946]: I1203 08:58:47.744730 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xd8h" event={"ID":"780e9ab7-484c-489d-8594-14f42bd8d8f8","Type":"ContainerDied","Data":"0273836849c25c7e2bc31121d9b71d28fd05a7a9734b1afecbd39dfaddad2643"} Dec 03 08:58:48 crc kubenswrapper[4946]: I1203 08:58:48.758547 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xd8h" event={"ID":"780e9ab7-484c-489d-8594-14f42bd8d8f8","Type":"ContainerStarted","Data":"3b08b6d0459514bfd2e951736f57c483ed262fbbc38cdae422a7df64e5472e82"} Dec 03 08:58:48 crc kubenswrapper[4946]: I1203 08:58:48.787408 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2xd8h" podStartSLOduration=3.196293412 podStartE2EDuration="5.787380659s" podCreationTimestamp="2025-12-03 08:58:43 +0000 UTC" firstStartedPulling="2025-12-03 08:58:45.723642531 +0000 UTC m=+7718.520332650" lastFinishedPulling="2025-12-03 08:58:48.314729768 +0000 UTC m=+7721.111419897" observedRunningTime="2025-12-03 08:58:48.786812095 +0000 UTC m=+7721.583502204" watchObservedRunningTime="2025-12-03 08:58:48.787380659 +0000 UTC m=+7721.584070778" Dec 03 08:58:54 crc kubenswrapper[4946]: I1203 08:58:54.326388 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:54 crc kubenswrapper[4946]: I1203 08:58:54.327054 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:54 crc kubenswrapper[4946]: I1203 08:58:54.393669 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:54 crc kubenswrapper[4946]: I1203 08:58:54.868265 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:54 crc kubenswrapper[4946]: I1203 08:58:54.935354 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2xd8h"] Dec 03 08:58:56 crc kubenswrapper[4946]: I1203 08:58:56.829906 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2xd8h" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerName="registry-server" containerID="cri-o://3b08b6d0459514bfd2e951736f57c483ed262fbbc38cdae422a7df64e5472e82" gracePeriod=2 Dec 03 08:58:57 crc kubenswrapper[4946]: I1203 08:58:57.844356 4946 generic.go:334] "Generic (PLEG): container finished" podID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerID="3b08b6d0459514bfd2e951736f57c483ed262fbbc38cdae422a7df64e5472e82" exitCode=0 Dec 03 08:58:57 crc kubenswrapper[4946]: I1203 08:58:57.844453 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xd8h" event={"ID":"780e9ab7-484c-489d-8594-14f42bd8d8f8","Type":"ContainerDied","Data":"3b08b6d0459514bfd2e951736f57c483ed262fbbc38cdae422a7df64e5472e82"} Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.381729 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.503986 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-utilities\") pod \"780e9ab7-484c-489d-8594-14f42bd8d8f8\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.504118 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-catalog-content\") pod \"780e9ab7-484c-489d-8594-14f42bd8d8f8\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.504218 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss9t5\" (UniqueName: \"kubernetes.io/projected/780e9ab7-484c-489d-8594-14f42bd8d8f8-kube-api-access-ss9t5\") pod \"780e9ab7-484c-489d-8594-14f42bd8d8f8\" (UID: \"780e9ab7-484c-489d-8594-14f42bd8d8f8\") " Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.505416 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-utilities" (OuterVolumeSpecName: "utilities") pod "780e9ab7-484c-489d-8594-14f42bd8d8f8" (UID: "780e9ab7-484c-489d-8594-14f42bd8d8f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.514634 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/780e9ab7-484c-489d-8594-14f42bd8d8f8-kube-api-access-ss9t5" (OuterVolumeSpecName: "kube-api-access-ss9t5") pod "780e9ab7-484c-489d-8594-14f42bd8d8f8" (UID: "780e9ab7-484c-489d-8594-14f42bd8d8f8"). InnerVolumeSpecName "kube-api-access-ss9t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.570796 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "780e9ab7-484c-489d-8594-14f42bd8d8f8" (UID: "780e9ab7-484c-489d-8594-14f42bd8d8f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.606571 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.606604 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/780e9ab7-484c-489d-8594-14f42bd8d8f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.606618 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss9t5\" (UniqueName: \"kubernetes.io/projected/780e9ab7-484c-489d-8594-14f42bd8d8f8-kube-api-access-ss9t5\") on node \"crc\" DevicePath \"\"" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.861320 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xd8h" event={"ID":"780e9ab7-484c-489d-8594-14f42bd8d8f8","Type":"ContainerDied","Data":"2cd6a8001027157e5f91946cc10ab16430fa054411e3a2ca6ff9334ca7aa4914"} Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.861386 4946 scope.go:117] "RemoveContainer" containerID="3b08b6d0459514bfd2e951736f57c483ed262fbbc38cdae422a7df64e5472e82" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.861395 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2xd8h" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.890219 4946 scope.go:117] "RemoveContainer" containerID="0273836849c25c7e2bc31121d9b71d28fd05a7a9734b1afecbd39dfaddad2643" Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.907892 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2xd8h"] Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.913441 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2xd8h"] Dec 03 08:58:58 crc kubenswrapper[4946]: I1203 08:58:58.924214 4946 scope.go:117] "RemoveContainer" containerID="2e1bfd444f70df7cf550e6e0c8074f3f02a8d60d17631b523cfc5834a1e63373" Dec 03 08:58:59 crc kubenswrapper[4946]: I1203 08:58:59.601519 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" path="/var/lib/kubelet/pods/780e9ab7-484c-489d-8594-14f42bd8d8f8/volumes" Dec 03 08:59:23 crc kubenswrapper[4946]: I1203 08:59:23.039037 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:59:23 crc kubenswrapper[4946]: I1203 08:59:23.039773 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.038988 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f69v2"] Dec 03 08:59:42 crc kubenswrapper[4946]: E1203 08:59:42.039801 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerName="extract-utilities" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.039817 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerName="extract-utilities" Dec 03 08:59:42 crc kubenswrapper[4946]: E1203 08:59:42.039828 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerName="registry-server" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.039836 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerName="registry-server" Dec 03 08:59:42 crc kubenswrapper[4946]: E1203 08:59:42.039863 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerName="extract-content" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.039872 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerName="extract-content" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.040086 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="780e9ab7-484c-489d-8594-14f42bd8d8f8" containerName="registry-server" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.041297 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.062945 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f69v2"] Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.209507 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4hzt\" (UniqueName: \"kubernetes.io/projected/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-kube-api-access-c4hzt\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.209555 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-utilities\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.209603 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-catalog-content\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.311674 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4hzt\" (UniqueName: \"kubernetes.io/projected/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-kube-api-access-c4hzt\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.311774 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-utilities\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.311876 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-catalog-content\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.312390 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-utilities\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.312556 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-catalog-content\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.333843 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4hzt\" (UniqueName: \"kubernetes.io/projected/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-kube-api-access-c4hzt\") pod \"redhat-operators-f69v2\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.424685 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:42 crc kubenswrapper[4946]: I1203 08:59:42.891246 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f69v2"] Dec 03 08:59:43 crc kubenswrapper[4946]: I1203 08:59:43.341724 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f69v2" event={"ID":"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f","Type":"ContainerStarted","Data":"f618e45e5b6d161183f2d6c31de37c6fc837224d8fc674ce4c3be1f35007ce7b"} Dec 03 08:59:44 crc kubenswrapper[4946]: I1203 08:59:44.352074 4946 generic.go:334] "Generic (PLEG): container finished" podID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerID="423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33" exitCode=0 Dec 03 08:59:44 crc kubenswrapper[4946]: I1203 08:59:44.352206 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f69v2" event={"ID":"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f","Type":"ContainerDied","Data":"423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33"} Dec 03 08:59:45 crc kubenswrapper[4946]: I1203 08:59:45.363563 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f69v2" event={"ID":"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f","Type":"ContainerStarted","Data":"d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019"} Dec 03 08:59:46 crc kubenswrapper[4946]: I1203 08:59:46.373517 4946 generic.go:334] "Generic (PLEG): container finished" podID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerID="d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019" exitCode=0 Dec 03 08:59:46 crc kubenswrapper[4946]: I1203 08:59:46.373610 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f69v2" event={"ID":"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f","Type":"ContainerDied","Data":"d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019"} Dec 03 08:59:47 crc kubenswrapper[4946]: I1203 08:59:47.384767 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f69v2" event={"ID":"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f","Type":"ContainerStarted","Data":"480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3"} Dec 03 08:59:47 crc kubenswrapper[4946]: I1203 08:59:47.422435 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f69v2" podStartSLOduration=2.680197772 podStartE2EDuration="5.42240709s" podCreationTimestamp="2025-12-03 08:59:42 +0000 UTC" firstStartedPulling="2025-12-03 08:59:44.353343923 +0000 UTC m=+7777.150034032" lastFinishedPulling="2025-12-03 08:59:47.095553231 +0000 UTC m=+7779.892243350" observedRunningTime="2025-12-03 08:59:47.413583739 +0000 UTC m=+7780.210273878" watchObservedRunningTime="2025-12-03 08:59:47.42240709 +0000 UTC m=+7780.219097209" Dec 03 08:59:52 crc kubenswrapper[4946]: I1203 08:59:52.425821 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:52 crc kubenswrapper[4946]: I1203 08:59:52.427730 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 08:59:53 crc kubenswrapper[4946]: I1203 08:59:53.039512 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:59:53 crc kubenswrapper[4946]: I1203 08:59:53.039619 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:59:53 crc kubenswrapper[4946]: I1203 08:59:53.478269 4946 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f69v2" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="registry-server" probeResult="failure" output=< Dec 03 08:59:53 crc kubenswrapper[4946]: timeout: failed to connect service ":50051" within 1s Dec 03 08:59:53 crc kubenswrapper[4946]: > Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.216390 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb"] Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.218136 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.221810 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.222003 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.228136 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb"] Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.278078 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cefda6a-fa0d-49e4-8314-01ea8de17532-secret-volume\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.278154 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x6db\" (UniqueName: \"kubernetes.io/projected/0cefda6a-fa0d-49e4-8314-01ea8de17532-kube-api-access-9x6db\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.278178 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cefda6a-fa0d-49e4-8314-01ea8de17532-config-volume\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.379720 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cefda6a-fa0d-49e4-8314-01ea8de17532-config-volume\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.380120 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cefda6a-fa0d-49e4-8314-01ea8de17532-secret-volume\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.380165 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x6db\" (UniqueName: \"kubernetes.io/projected/0cefda6a-fa0d-49e4-8314-01ea8de17532-kube-api-access-9x6db\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.380735 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cefda6a-fa0d-49e4-8314-01ea8de17532-config-volume\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.387673 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cefda6a-fa0d-49e4-8314-01ea8de17532-secret-volume\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.396544 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x6db\" (UniqueName: \"kubernetes.io/projected/0cefda6a-fa0d-49e4-8314-01ea8de17532-kube-api-access-9x6db\") pod \"collect-profiles-29412540-hj6vb\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.540129 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:00 crc kubenswrapper[4946]: I1203 09:00:00.996565 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb"] Dec 03 09:00:01 crc kubenswrapper[4946]: W1203 09:00:01.011061 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cefda6a_fa0d_49e4_8314_01ea8de17532.slice/crio-1c35c128605e9cd294b2f8aad3f07c83fee7780f588f39def200884758ea7bd7 WatchSource:0}: Error finding container 1c35c128605e9cd294b2f8aad3f07c83fee7780f588f39def200884758ea7bd7: Status 404 returned error can't find the container with id 1c35c128605e9cd294b2f8aad3f07c83fee7780f588f39def200884758ea7bd7 Dec 03 09:00:01 crc kubenswrapper[4946]: I1203 09:00:01.517427 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" event={"ID":"0cefda6a-fa0d-49e4-8314-01ea8de17532","Type":"ContainerStarted","Data":"1c35c128605e9cd294b2f8aad3f07c83fee7780f588f39def200884758ea7bd7"} Dec 03 09:00:02 crc kubenswrapper[4946]: I1203 09:00:02.464542 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 09:00:02 crc kubenswrapper[4946]: I1203 09:00:02.527600 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 09:00:02 crc kubenswrapper[4946]: I1203 09:00:02.527638 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" event={"ID":"0cefda6a-fa0d-49e4-8314-01ea8de17532","Type":"ContainerDied","Data":"78cea21c5861563ec5b88eb7b630cb4aa99a69bab6cde2687fad4cceda2c85fb"} Dec 03 09:00:02 crc kubenswrapper[4946]: I1203 09:00:02.527602 4946 generic.go:334] "Generic (PLEG): container finished" podID="0cefda6a-fa0d-49e4-8314-01ea8de17532" containerID="78cea21c5861563ec5b88eb7b630cb4aa99a69bab6cde2687fad4cceda2c85fb" exitCode=0 Dec 03 09:00:02 crc kubenswrapper[4946]: I1203 09:00:02.696756 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f69v2"] Dec 03 09:00:03 crc kubenswrapper[4946]: I1203 09:00:03.535636 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f69v2" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="registry-server" containerID="cri-o://480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3" gracePeriod=2 Dec 03 09:00:03 crc kubenswrapper[4946]: I1203 09:00:03.898712 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:03 crc kubenswrapper[4946]: I1203 09:00:03.979639 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.067042 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4hzt\" (UniqueName: \"kubernetes.io/projected/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-kube-api-access-c4hzt\") pod \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.067115 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cefda6a-fa0d-49e4-8314-01ea8de17532-secret-volume\") pod \"0cefda6a-fa0d-49e4-8314-01ea8de17532\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.067179 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-catalog-content\") pod \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.067241 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x6db\" (UniqueName: \"kubernetes.io/projected/0cefda6a-fa0d-49e4-8314-01ea8de17532-kube-api-access-9x6db\") pod \"0cefda6a-fa0d-49e4-8314-01ea8de17532\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.067323 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-utilities\") pod \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\" (UID: \"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f\") " Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.068251 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-utilities" (OuterVolumeSpecName: "utilities") pod "d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" (UID: "d1cd8d04-1501-47eb-95a8-98fd42a8aa6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.068696 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cefda6a-fa0d-49e4-8314-01ea8de17532-config-volume\") pod \"0cefda6a-fa0d-49e4-8314-01ea8de17532\" (UID: \"0cefda6a-fa0d-49e4-8314-01ea8de17532\") " Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.069090 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cefda6a-fa0d-49e4-8314-01ea8de17532-config-volume" (OuterVolumeSpecName: "config-volume") pod "0cefda6a-fa0d-49e4-8314-01ea8de17532" (UID: "0cefda6a-fa0d-49e4-8314-01ea8de17532"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.069273 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.069305 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cefda6a-fa0d-49e4-8314-01ea8de17532-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.072894 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-kube-api-access-c4hzt" (OuterVolumeSpecName: "kube-api-access-c4hzt") pod "d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" (UID: "d1cd8d04-1501-47eb-95a8-98fd42a8aa6f"). InnerVolumeSpecName "kube-api-access-c4hzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.073006 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cefda6a-fa0d-49e4-8314-01ea8de17532-kube-api-access-9x6db" (OuterVolumeSpecName: "kube-api-access-9x6db") pod "0cefda6a-fa0d-49e4-8314-01ea8de17532" (UID: "0cefda6a-fa0d-49e4-8314-01ea8de17532"). InnerVolumeSpecName "kube-api-access-9x6db". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.074867 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cefda6a-fa0d-49e4-8314-01ea8de17532-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0cefda6a-fa0d-49e4-8314-01ea8de17532" (UID: "0cefda6a-fa0d-49e4-8314-01ea8de17532"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.170626 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x6db\" (UniqueName: \"kubernetes.io/projected/0cefda6a-fa0d-49e4-8314-01ea8de17532-kube-api-access-9x6db\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.170663 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4hzt\" (UniqueName: \"kubernetes.io/projected/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-kube-api-access-c4hzt\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.170676 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cefda6a-fa0d-49e4-8314-01ea8de17532-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.189568 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" (UID: "d1cd8d04-1501-47eb-95a8-98fd42a8aa6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.272862 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.546921 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" event={"ID":"0cefda6a-fa0d-49e4-8314-01ea8de17532","Type":"ContainerDied","Data":"1c35c128605e9cd294b2f8aad3f07c83fee7780f588f39def200884758ea7bd7"} Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.547029 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c35c128605e9cd294b2f8aad3f07c83fee7780f588f39def200884758ea7bd7" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.547183 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-hj6vb" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.550092 4946 generic.go:334] "Generic (PLEG): container finished" podID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerID="480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3" exitCode=0 Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.550161 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f69v2" event={"ID":"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f","Type":"ContainerDied","Data":"480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3"} Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.550180 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f69v2" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.550220 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f69v2" event={"ID":"d1cd8d04-1501-47eb-95a8-98fd42a8aa6f","Type":"ContainerDied","Data":"f618e45e5b6d161183f2d6c31de37c6fc837224d8fc674ce4c3be1f35007ce7b"} Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.550265 4946 scope.go:117] "RemoveContainer" containerID="480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.591029 4946 scope.go:117] "RemoveContainer" containerID="d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.607602 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f69v2"] Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.620460 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f69v2"] Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.630544 4946 scope.go:117] "RemoveContainer" containerID="423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.655674 4946 scope.go:117] "RemoveContainer" containerID="480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3" Dec 03 09:00:04 crc kubenswrapper[4946]: E1203 09:00:04.656345 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3\": container with ID starting with 480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3 not found: ID does not exist" containerID="480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.656408 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3"} err="failed to get container status \"480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3\": rpc error: code = NotFound desc = could not find container \"480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3\": container with ID starting with 480a72c4a8ae5a506db02dfd32f4921867605aa3c03356f053a23de5e0a55af3 not found: ID does not exist" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.656444 4946 scope.go:117] "RemoveContainer" containerID="d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019" Dec 03 09:00:04 crc kubenswrapper[4946]: E1203 09:00:04.657123 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019\": container with ID starting with d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019 not found: ID does not exist" containerID="d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.657163 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019"} err="failed to get container status \"d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019\": rpc error: code = NotFound desc = could not find container \"d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019\": container with ID starting with d380e17d7bd983ab2093f617b0f3241637ac97fb0b2081b745307f0695c43019 not found: ID does not exist" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.657191 4946 scope.go:117] "RemoveContainer" containerID="423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33" Dec 03 09:00:04 crc kubenswrapper[4946]: E1203 09:00:04.657932 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33\": container with ID starting with 423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33 not found: ID does not exist" containerID="423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.657959 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33"} err="failed to get container status \"423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33\": rpc error: code = NotFound desc = could not find container \"423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33\": container with ID starting with 423a0684f8b25fbdcd75beebedce937618e95825017e737e3982720ddf0fed33 not found: ID does not exist" Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.992205 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4"] Dec 03 09:00:04 crc kubenswrapper[4946]: I1203 09:00:04.999506 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412495-xnwg4"] Dec 03 09:00:05 crc kubenswrapper[4946]: I1203 09:00:05.610093 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe" path="/var/lib/kubelet/pods/8ad46f4f-0838-4cf6-80e4-dbc2361f3bfe/volumes" Dec 03 09:00:05 crc kubenswrapper[4946]: I1203 09:00:05.611315 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" path="/var/lib/kubelet/pods/d1cd8d04-1501-47eb-95a8-98fd42a8aa6f/volumes" Dec 03 09:00:15 crc kubenswrapper[4946]: I1203 09:00:15.521478 4946 scope.go:117] "RemoveContainer" containerID="6a75dd080ac098440b8fad4543dfebd6d569a0265d2431ebcf2b13b3f8ec4a3b" Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.039923 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.040829 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.040905 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.042033 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.042140 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" gracePeriod=600 Dec 03 09:00:23 crc kubenswrapper[4946]: E1203 09:00:23.187223 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.740030 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" exitCode=0 Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.740097 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6"} Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.740147 4946 scope.go:117] "RemoveContainer" containerID="6ec43767614acc0b2c9ae69b6756fbf948a7fd22b3606e6c66341c4c8adda2fe" Dec 03 09:00:23 crc kubenswrapper[4946]: I1203 09:00:23.740826 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:00:23 crc kubenswrapper[4946]: E1203 09:00:23.741102 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:00:36 crc kubenswrapper[4946]: I1203 09:00:36.594634 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:00:36 crc kubenswrapper[4946]: E1203 09:00:36.595936 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:00:47 crc kubenswrapper[4946]: I1203 09:00:47.601308 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:00:47 crc kubenswrapper[4946]: E1203 09:00:47.603220 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:00:58 crc kubenswrapper[4946]: I1203 09:00:58.593446 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:00:58 crc kubenswrapper[4946]: E1203 09:00:58.594548 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:01:11 crc kubenswrapper[4946]: I1203 09:01:11.592969 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:01:11 crc kubenswrapper[4946]: E1203 09:01:11.594882 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:01:24 crc kubenswrapper[4946]: I1203 09:01:24.593875 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:01:24 crc kubenswrapper[4946]: E1203 09:01:24.595113 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:01:40 crc kubenswrapper[4946]: I1203 09:01:40.593634 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:01:40 crc kubenswrapper[4946]: E1203 09:01:40.594395 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:01:53 crc kubenswrapper[4946]: I1203 09:01:53.593540 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:01:53 crc kubenswrapper[4946]: E1203 09:01:53.596314 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:02:06 crc kubenswrapper[4946]: I1203 09:02:06.593265 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:02:06 crc kubenswrapper[4946]: E1203 09:02:06.594537 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:02:18 crc kubenswrapper[4946]: I1203 09:02:18.593479 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:02:18 crc kubenswrapper[4946]: E1203 09:02:18.594928 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:02:31 crc kubenswrapper[4946]: I1203 09:02:31.593847 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:02:31 crc kubenswrapper[4946]: E1203 09:02:31.596113 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:02:45 crc kubenswrapper[4946]: I1203 09:02:45.593027 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:02:45 crc kubenswrapper[4946]: E1203 09:02:45.594232 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:02:58 crc kubenswrapper[4946]: I1203 09:02:58.592923 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:02:58 crc kubenswrapper[4946]: E1203 09:02:58.594054 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:03:12 crc kubenswrapper[4946]: I1203 09:03:12.593470 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:03:12 crc kubenswrapper[4946]: E1203 09:03:12.594179 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:03:27 crc kubenswrapper[4946]: I1203 09:03:27.600965 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:03:27 crc kubenswrapper[4946]: E1203 09:03:27.601757 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:03:38 crc kubenswrapper[4946]: I1203 09:03:38.593358 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:03:38 crc kubenswrapper[4946]: E1203 09:03:38.594553 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:03:50 crc kubenswrapper[4946]: I1203 09:03:50.593791 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:03:50 crc kubenswrapper[4946]: E1203 09:03:50.595287 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:04:01 crc kubenswrapper[4946]: I1203 09:04:01.593412 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:04:01 crc kubenswrapper[4946]: E1203 09:04:01.594546 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:04:14 crc kubenswrapper[4946]: I1203 09:04:14.593155 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:04:14 crc kubenswrapper[4946]: E1203 09:04:14.593989 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:04:27 crc kubenswrapper[4946]: I1203 09:04:27.598888 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:04:27 crc kubenswrapper[4946]: E1203 09:04:27.599787 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:04:39 crc kubenswrapper[4946]: I1203 09:04:39.592795 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:04:39 crc kubenswrapper[4946]: E1203 09:04:39.593520 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:04:53 crc kubenswrapper[4946]: I1203 09:04:53.593636 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:04:53 crc kubenswrapper[4946]: E1203 09:04:53.594847 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:05:05 crc kubenswrapper[4946]: I1203 09:05:05.593353 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:05:05 crc kubenswrapper[4946]: E1203 09:05:05.594063 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:05:17 crc kubenswrapper[4946]: I1203 09:05:17.601010 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:05:17 crc kubenswrapper[4946]: E1203 09:05:17.602094 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:05:28 crc kubenswrapper[4946]: I1203 09:05:28.594086 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:05:29 crc kubenswrapper[4946]: I1203 09:05:29.765245 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"8a3899d0e6f82103ad9642f8af450b7d00441adf16c0a905df352bdafc37da27"} Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.162498 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jt4fv"] Dec 03 09:07:10 crc kubenswrapper[4946]: E1203 09:07:10.163321 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="extract-content" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.163334 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="extract-content" Dec 03 09:07:10 crc kubenswrapper[4946]: E1203 09:07:10.163354 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cefda6a-fa0d-49e4-8314-01ea8de17532" containerName="collect-profiles" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.163360 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cefda6a-fa0d-49e4-8314-01ea8de17532" containerName="collect-profiles" Dec 03 09:07:10 crc kubenswrapper[4946]: E1203 09:07:10.163377 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="extract-utilities" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.163384 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="extract-utilities" Dec 03 09:07:10 crc kubenswrapper[4946]: E1203 09:07:10.163399 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="registry-server" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.163405 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="registry-server" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.163522 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1cd8d04-1501-47eb-95a8-98fd42a8aa6f" containerName="registry-server" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.163535 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cefda6a-fa0d-49e4-8314-01ea8de17532" containerName="collect-profiles" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.169019 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.185315 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf7hp\" (UniqueName: \"kubernetes.io/projected/25399b85-a422-4afd-9f30-995f24eeb4cc-kube-api-access-gf7hp\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.185408 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jt4fv"] Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.185566 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-catalog-content\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.185681 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-utilities\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.286961 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-utilities\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.287018 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf7hp\" (UniqueName: \"kubernetes.io/projected/25399b85-a422-4afd-9f30-995f24eeb4cc-kube-api-access-gf7hp\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.287086 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-catalog-content\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.287586 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-utilities\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.287635 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-catalog-content\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.313443 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf7hp\" (UniqueName: \"kubernetes.io/projected/25399b85-a422-4afd-9f30-995f24eeb4cc-kube-api-access-gf7hp\") pod \"redhat-marketplace-jt4fv\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.491081 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:10 crc kubenswrapper[4946]: I1203 09:07:10.930206 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jt4fv"] Dec 03 09:07:11 crc kubenswrapper[4946]: I1203 09:07:11.687500 4946 generic.go:334] "Generic (PLEG): container finished" podID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerID="d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b" exitCode=0 Dec 03 09:07:11 crc kubenswrapper[4946]: I1203 09:07:11.687899 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jt4fv" event={"ID":"25399b85-a422-4afd-9f30-995f24eeb4cc","Type":"ContainerDied","Data":"d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b"} Dec 03 09:07:11 crc kubenswrapper[4946]: I1203 09:07:11.687944 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jt4fv" event={"ID":"25399b85-a422-4afd-9f30-995f24eeb4cc","Type":"ContainerStarted","Data":"d8934a137a80cfd92e578f5b9b48dc815e75906ea111b6fd11a224f4568b9bcb"} Dec 03 09:07:11 crc kubenswrapper[4946]: I1203 09:07:11.690601 4946 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:07:13 crc kubenswrapper[4946]: I1203 09:07:13.709965 4946 generic.go:334] "Generic (PLEG): container finished" podID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerID="b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932" exitCode=0 Dec 03 09:07:13 crc kubenswrapper[4946]: I1203 09:07:13.710058 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jt4fv" event={"ID":"25399b85-a422-4afd-9f30-995f24eeb4cc","Type":"ContainerDied","Data":"b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932"} Dec 03 09:07:14 crc kubenswrapper[4946]: I1203 09:07:14.724075 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jt4fv" event={"ID":"25399b85-a422-4afd-9f30-995f24eeb4cc","Type":"ContainerStarted","Data":"d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb"} Dec 03 09:07:14 crc kubenswrapper[4946]: I1203 09:07:14.760371 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jt4fv" podStartSLOduration=2.230207036 podStartE2EDuration="4.76028064s" podCreationTimestamp="2025-12-03 09:07:10 +0000 UTC" firstStartedPulling="2025-12-03 09:07:11.690220569 +0000 UTC m=+8224.486910698" lastFinishedPulling="2025-12-03 09:07:14.220294183 +0000 UTC m=+8227.016984302" observedRunningTime="2025-12-03 09:07:14.752401856 +0000 UTC m=+8227.549092015" watchObservedRunningTime="2025-12-03 09:07:14.76028064 +0000 UTC m=+8227.556970749" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.543257 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-blkjd"] Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.546715 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.556991 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blkjd"] Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.686359 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-catalog-content\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.686417 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dlgq\" (UniqueName: \"kubernetes.io/projected/7b88bb54-b257-486a-b5a3-458a942db292-kube-api-access-8dlgq\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.686442 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-utilities\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.787863 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-catalog-content\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.788128 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dlgq\" (UniqueName: \"kubernetes.io/projected/7b88bb54-b257-486a-b5a3-458a942db292-kube-api-access-8dlgq\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.788206 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-utilities\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.788586 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-catalog-content\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.788665 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-utilities\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.811025 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dlgq\" (UniqueName: \"kubernetes.io/projected/7b88bb54-b257-486a-b5a3-458a942db292-kube-api-access-8dlgq\") pod \"certified-operators-blkjd\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:16 crc kubenswrapper[4946]: I1203 09:07:16.870421 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:17 crc kubenswrapper[4946]: I1203 09:07:17.149091 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blkjd"] Dec 03 09:07:17 crc kubenswrapper[4946]: W1203 09:07:17.153050 4946 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b88bb54_b257_486a_b5a3_458a942db292.slice/crio-1e32b2dde8a068910c4fc019d25b8a57ec3bf39802a232258b8d5194f1cdb414 WatchSource:0}: Error finding container 1e32b2dde8a068910c4fc019d25b8a57ec3bf39802a232258b8d5194f1cdb414: Status 404 returned error can't find the container with id 1e32b2dde8a068910c4fc019d25b8a57ec3bf39802a232258b8d5194f1cdb414 Dec 03 09:07:17 crc kubenswrapper[4946]: I1203 09:07:17.748429 4946 generic.go:334] "Generic (PLEG): container finished" podID="7b88bb54-b257-486a-b5a3-458a942db292" containerID="7973f82f0eb496c82085134bb45872514afdca99aa6d145564551326a3e6248e" exitCode=0 Dec 03 09:07:17 crc kubenswrapper[4946]: I1203 09:07:17.748577 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blkjd" event={"ID":"7b88bb54-b257-486a-b5a3-458a942db292","Type":"ContainerDied","Data":"7973f82f0eb496c82085134bb45872514afdca99aa6d145564551326a3e6248e"} Dec 03 09:07:17 crc kubenswrapper[4946]: I1203 09:07:17.748706 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blkjd" event={"ID":"7b88bb54-b257-486a-b5a3-458a942db292","Type":"ContainerStarted","Data":"1e32b2dde8a068910c4fc019d25b8a57ec3bf39802a232258b8d5194f1cdb414"} Dec 03 09:07:19 crc kubenswrapper[4946]: I1203 09:07:19.767836 4946 generic.go:334] "Generic (PLEG): container finished" podID="7b88bb54-b257-486a-b5a3-458a942db292" containerID="a996ae39cc695f41b404247f1b2dc9f6cb67ba7a1809e343adaab5367089140c" exitCode=0 Dec 03 09:07:19 crc kubenswrapper[4946]: I1203 09:07:19.767893 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blkjd" event={"ID":"7b88bb54-b257-486a-b5a3-458a942db292","Type":"ContainerDied","Data":"a996ae39cc695f41b404247f1b2dc9f6cb67ba7a1809e343adaab5367089140c"} Dec 03 09:07:20 crc kubenswrapper[4946]: I1203 09:07:20.491903 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:20 crc kubenswrapper[4946]: I1203 09:07:20.492348 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:20 crc kubenswrapper[4946]: I1203 09:07:20.577699 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:20 crc kubenswrapper[4946]: I1203 09:07:20.779511 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blkjd" event={"ID":"7b88bb54-b257-486a-b5a3-458a942db292","Type":"ContainerStarted","Data":"61082f909ac1f15bde9b690e17c86b81756526f5578593ddd44e1e295e24e414"} Dec 03 09:07:20 crc kubenswrapper[4946]: I1203 09:07:20.805543 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-blkjd" podStartSLOduration=2.361278122 podStartE2EDuration="4.805527114s" podCreationTimestamp="2025-12-03 09:07:16 +0000 UTC" firstStartedPulling="2025-12-03 09:07:17.751118318 +0000 UTC m=+8230.547808437" lastFinishedPulling="2025-12-03 09:07:20.19536731 +0000 UTC m=+8232.992057429" observedRunningTime="2025-12-03 09:07:20.802586344 +0000 UTC m=+8233.599276453" watchObservedRunningTime="2025-12-03 09:07:20.805527114 +0000 UTC m=+8233.602217223" Dec 03 09:07:20 crc kubenswrapper[4946]: I1203 09:07:20.852132 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:22 crc kubenswrapper[4946]: I1203 09:07:22.530994 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jt4fv"] Dec 03 09:07:22 crc kubenswrapper[4946]: I1203 09:07:22.798987 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jt4fv" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerName="registry-server" containerID="cri-o://d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb" gracePeriod=2 Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.269517 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.407308 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf7hp\" (UniqueName: \"kubernetes.io/projected/25399b85-a422-4afd-9f30-995f24eeb4cc-kube-api-access-gf7hp\") pod \"25399b85-a422-4afd-9f30-995f24eeb4cc\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.407370 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-utilities\") pod \"25399b85-a422-4afd-9f30-995f24eeb4cc\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.407449 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-catalog-content\") pod \"25399b85-a422-4afd-9f30-995f24eeb4cc\" (UID: \"25399b85-a422-4afd-9f30-995f24eeb4cc\") " Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.408441 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-utilities" (OuterVolumeSpecName: "utilities") pod "25399b85-a422-4afd-9f30-995f24eeb4cc" (UID: "25399b85-a422-4afd-9f30-995f24eeb4cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.413494 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25399b85-a422-4afd-9f30-995f24eeb4cc-kube-api-access-gf7hp" (OuterVolumeSpecName: "kube-api-access-gf7hp") pod "25399b85-a422-4afd-9f30-995f24eeb4cc" (UID: "25399b85-a422-4afd-9f30-995f24eeb4cc"). InnerVolumeSpecName "kube-api-access-gf7hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.445734 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25399b85-a422-4afd-9f30-995f24eeb4cc" (UID: "25399b85-a422-4afd-9f30-995f24eeb4cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.509250 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.509797 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf7hp\" (UniqueName: \"kubernetes.io/projected/25399b85-a422-4afd-9f30-995f24eeb4cc-kube-api-access-gf7hp\") on node \"crc\" DevicePath \"\"" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.509837 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25399b85-a422-4afd-9f30-995f24eeb4cc-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.807679 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jt4fv" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.807592 4946 generic.go:334] "Generic (PLEG): container finished" podID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerID="d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb" exitCode=0 Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.807724 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jt4fv" event={"ID":"25399b85-a422-4afd-9f30-995f24eeb4cc","Type":"ContainerDied","Data":"d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb"} Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.809031 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jt4fv" event={"ID":"25399b85-a422-4afd-9f30-995f24eeb4cc","Type":"ContainerDied","Data":"d8934a137a80cfd92e578f5b9b48dc815e75906ea111b6fd11a224f4568b9bcb"} Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.809148 4946 scope.go:117] "RemoveContainer" containerID="d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.835536 4946 scope.go:117] "RemoveContainer" containerID="b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.843954 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jt4fv"] Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.852932 4946 scope.go:117] "RemoveContainer" containerID="d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.855221 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jt4fv"] Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.900389 4946 scope.go:117] "RemoveContainer" containerID="d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb" Dec 03 09:07:23 crc kubenswrapper[4946]: E1203 09:07:23.900845 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb\": container with ID starting with d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb not found: ID does not exist" containerID="d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.900900 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb"} err="failed to get container status \"d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb\": rpc error: code = NotFound desc = could not find container \"d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb\": container with ID starting with d678b6f407a866960a664ecddeb3c8b7635dd5011f0e197192e8c1565c2edefb not found: ID does not exist" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.900926 4946 scope.go:117] "RemoveContainer" containerID="b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932" Dec 03 09:07:23 crc kubenswrapper[4946]: E1203 09:07:23.901240 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932\": container with ID starting with b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932 not found: ID does not exist" containerID="b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.901266 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932"} err="failed to get container status \"b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932\": rpc error: code = NotFound desc = could not find container \"b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932\": container with ID starting with b3f4b0aefc96f4255ebd84effcb7abea38830194caa58d8d0810cf7a5e542932 not found: ID does not exist" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.901284 4946 scope.go:117] "RemoveContainer" containerID="d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b" Dec 03 09:07:23 crc kubenswrapper[4946]: E1203 09:07:23.901541 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b\": container with ID starting with d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b not found: ID does not exist" containerID="d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b" Dec 03 09:07:23 crc kubenswrapper[4946]: I1203 09:07:23.901579 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b"} err="failed to get container status \"d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b\": rpc error: code = NotFound desc = could not find container \"d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b\": container with ID starting with d5414e40df90c326f46feb656286139c049179a622b695f3a37815a4c79ade4b not found: ID does not exist" Dec 03 09:07:25 crc kubenswrapper[4946]: I1203 09:07:25.609868 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" path="/var/lib/kubelet/pods/25399b85-a422-4afd-9f30-995f24eeb4cc/volumes" Dec 03 09:07:26 crc kubenswrapper[4946]: I1203 09:07:26.870821 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:26 crc kubenswrapper[4946]: I1203 09:07:26.870877 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:26 crc kubenswrapper[4946]: I1203 09:07:26.947178 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:27 crc kubenswrapper[4946]: I1203 09:07:27.923727 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:27 crc kubenswrapper[4946]: I1203 09:07:27.976675 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-blkjd"] Dec 03 09:07:29 crc kubenswrapper[4946]: I1203 09:07:29.862975 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-blkjd" podUID="7b88bb54-b257-486a-b5a3-458a942db292" containerName="registry-server" containerID="cri-o://61082f909ac1f15bde9b690e17c86b81756526f5578593ddd44e1e295e24e414" gracePeriod=2 Dec 03 09:07:30 crc kubenswrapper[4946]: I1203 09:07:30.875451 4946 generic.go:334] "Generic (PLEG): container finished" podID="7b88bb54-b257-486a-b5a3-458a942db292" containerID="61082f909ac1f15bde9b690e17c86b81756526f5578593ddd44e1e295e24e414" exitCode=0 Dec 03 09:07:30 crc kubenswrapper[4946]: I1203 09:07:30.875531 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blkjd" event={"ID":"7b88bb54-b257-486a-b5a3-458a942db292","Type":"ContainerDied","Data":"61082f909ac1f15bde9b690e17c86b81756526f5578593ddd44e1e295e24e414"} Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.435584 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.541461 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dlgq\" (UniqueName: \"kubernetes.io/projected/7b88bb54-b257-486a-b5a3-458a942db292-kube-api-access-8dlgq\") pod \"7b88bb54-b257-486a-b5a3-458a942db292\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.541532 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-utilities\") pod \"7b88bb54-b257-486a-b5a3-458a942db292\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.541628 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-catalog-content\") pod \"7b88bb54-b257-486a-b5a3-458a942db292\" (UID: \"7b88bb54-b257-486a-b5a3-458a942db292\") " Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.542450 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-utilities" (OuterVolumeSpecName: "utilities") pod "7b88bb54-b257-486a-b5a3-458a942db292" (UID: "7b88bb54-b257-486a-b5a3-458a942db292"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.554065 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b88bb54-b257-486a-b5a3-458a942db292-kube-api-access-8dlgq" (OuterVolumeSpecName: "kube-api-access-8dlgq") pod "7b88bb54-b257-486a-b5a3-458a942db292" (UID: "7b88bb54-b257-486a-b5a3-458a942db292"). InnerVolumeSpecName "kube-api-access-8dlgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.606299 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b88bb54-b257-486a-b5a3-458a942db292" (UID: "7b88bb54-b257-486a-b5a3-458a942db292"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.643591 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dlgq\" (UniqueName: \"kubernetes.io/projected/7b88bb54-b257-486a-b5a3-458a942db292-kube-api-access-8dlgq\") on node \"crc\" DevicePath \"\"" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.643787 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.644269 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b88bb54-b257-486a-b5a3-458a942db292-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.886567 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blkjd" event={"ID":"7b88bb54-b257-486a-b5a3-458a942db292","Type":"ContainerDied","Data":"1e32b2dde8a068910c4fc019d25b8a57ec3bf39802a232258b8d5194f1cdb414"} Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.886644 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blkjd" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.886924 4946 scope.go:117] "RemoveContainer" containerID="61082f909ac1f15bde9b690e17c86b81756526f5578593ddd44e1e295e24e414" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.911755 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-blkjd"] Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.916104 4946 scope.go:117] "RemoveContainer" containerID="a996ae39cc695f41b404247f1b2dc9f6cb67ba7a1809e343adaab5367089140c" Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.920116 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-blkjd"] Dec 03 09:07:31 crc kubenswrapper[4946]: I1203 09:07:31.943433 4946 scope.go:117] "RemoveContainer" containerID="7973f82f0eb496c82085134bb45872514afdca99aa6d145564551326a3e6248e" Dec 03 09:07:33 crc kubenswrapper[4946]: I1203 09:07:33.602689 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b88bb54-b257-486a-b5a3-458a942db292" path="/var/lib/kubelet/pods/7b88bb54-b257-486a-b5a3-458a942db292/volumes" Dec 03 09:07:53 crc kubenswrapper[4946]: I1203 09:07:53.039315 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:07:53 crc kubenswrapper[4946]: I1203 09:07:53.040930 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:08:23 crc kubenswrapper[4946]: I1203 09:08:23.039232 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:08:23 crc kubenswrapper[4946]: I1203 09:08:23.039861 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.918416 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fxqkt"] Dec 03 09:08:50 crc kubenswrapper[4946]: E1203 09:08:50.919671 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b88bb54-b257-486a-b5a3-458a942db292" containerName="extract-utilities" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.919694 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b88bb54-b257-486a-b5a3-458a942db292" containerName="extract-utilities" Dec 03 09:08:50 crc kubenswrapper[4946]: E1203 09:08:50.919726 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerName="extract-content" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.919764 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerName="extract-content" Dec 03 09:08:50 crc kubenswrapper[4946]: E1203 09:08:50.919788 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerName="extract-utilities" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.919800 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerName="extract-utilities" Dec 03 09:08:50 crc kubenswrapper[4946]: E1203 09:08:50.919818 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b88bb54-b257-486a-b5a3-458a942db292" containerName="extract-content" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.919828 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b88bb54-b257-486a-b5a3-458a942db292" containerName="extract-content" Dec 03 09:08:50 crc kubenswrapper[4946]: E1203 09:08:50.919848 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b88bb54-b257-486a-b5a3-458a942db292" containerName="registry-server" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.919858 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b88bb54-b257-486a-b5a3-458a942db292" containerName="registry-server" Dec 03 09:08:50 crc kubenswrapper[4946]: E1203 09:08:50.919877 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerName="registry-server" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.919904 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerName="registry-server" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.920133 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="25399b85-a422-4afd-9f30-995f24eeb4cc" containerName="registry-server" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.920169 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b88bb54-b257-486a-b5a3-458a942db292" containerName="registry-server" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.921805 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:50 crc kubenswrapper[4946]: I1203 09:08:50.933267 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fxqkt"] Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.031526 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klpqk\" (UniqueName: \"kubernetes.io/projected/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-kube-api-access-klpqk\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.031827 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-catalog-content\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.031994 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-utilities\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.133333 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klpqk\" (UniqueName: \"kubernetes.io/projected/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-kube-api-access-klpqk\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.133408 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-catalog-content\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.133449 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-utilities\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.134115 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-utilities\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.134166 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-catalog-content\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.158844 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klpqk\" (UniqueName: \"kubernetes.io/projected/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-kube-api-access-klpqk\") pod \"community-operators-fxqkt\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.247927 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:08:51 crc kubenswrapper[4946]: I1203 09:08:51.808213 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fxqkt"] Dec 03 09:08:52 crc kubenswrapper[4946]: I1203 09:08:52.663657 4946 generic.go:334] "Generic (PLEG): container finished" podID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerID="a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b" exitCode=0 Dec 03 09:08:52 crc kubenswrapper[4946]: I1203 09:08:52.663701 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxqkt" event={"ID":"6e13dce2-6b6f-4ffd-850a-9a40b522bbca","Type":"ContainerDied","Data":"a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b"} Dec 03 09:08:52 crc kubenswrapper[4946]: I1203 09:08:52.663726 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxqkt" event={"ID":"6e13dce2-6b6f-4ffd-850a-9a40b522bbca","Type":"ContainerStarted","Data":"02ff560f30b42a7f4c5bf9ff123dc28451531ed3d357018060664ff6c598190a"} Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.039470 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.039558 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.039627 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.040533 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a3899d0e6f82103ad9642f8af450b7d00441adf16c0a905df352bdafc37da27"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.040648 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://8a3899d0e6f82103ad9642f8af450b7d00441adf16c0a905df352bdafc37da27" gracePeriod=600 Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.678570 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxqkt" event={"ID":"6e13dce2-6b6f-4ffd-850a-9a40b522bbca","Type":"ContainerStarted","Data":"46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a"} Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.682050 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="8a3899d0e6f82103ad9642f8af450b7d00441adf16c0a905df352bdafc37da27" exitCode=0 Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.682102 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"8a3899d0e6f82103ad9642f8af450b7d00441adf16c0a905df352bdafc37da27"} Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.682145 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0"} Dec 03 09:08:53 crc kubenswrapper[4946]: I1203 09:08:53.682181 4946 scope.go:117] "RemoveContainer" containerID="e73e33fab09cbaf21a651213d38d11b3ca16c7d9887f71c2c801b4d020e078d6" Dec 03 09:08:54 crc kubenswrapper[4946]: I1203 09:08:54.702174 4946 generic.go:334] "Generic (PLEG): container finished" podID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerID="46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a" exitCode=0 Dec 03 09:08:54 crc kubenswrapper[4946]: I1203 09:08:54.702212 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxqkt" event={"ID":"6e13dce2-6b6f-4ffd-850a-9a40b522bbca","Type":"ContainerDied","Data":"46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a"} Dec 03 09:08:55 crc kubenswrapper[4946]: I1203 09:08:55.714302 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxqkt" event={"ID":"6e13dce2-6b6f-4ffd-850a-9a40b522bbca","Type":"ContainerStarted","Data":"17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82"} Dec 03 09:08:55 crc kubenswrapper[4946]: I1203 09:08:55.732666 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fxqkt" podStartSLOduration=3.07234455 podStartE2EDuration="5.732631224s" podCreationTimestamp="2025-12-03 09:08:50 +0000 UTC" firstStartedPulling="2025-12-03 09:08:52.666497489 +0000 UTC m=+8325.463187598" lastFinishedPulling="2025-12-03 09:08:55.326784163 +0000 UTC m=+8328.123474272" observedRunningTime="2025-12-03 09:08:55.72991469 +0000 UTC m=+8328.526604829" watchObservedRunningTime="2025-12-03 09:08:55.732631224 +0000 UTC m=+8328.529321343" Dec 03 09:09:01 crc kubenswrapper[4946]: I1203 09:09:01.248852 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:09:01 crc kubenswrapper[4946]: I1203 09:09:01.249488 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:09:01 crc kubenswrapper[4946]: I1203 09:09:01.305179 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:09:01 crc kubenswrapper[4946]: I1203 09:09:01.830423 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:09:01 crc kubenswrapper[4946]: I1203 09:09:01.878947 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fxqkt"] Dec 03 09:09:03 crc kubenswrapper[4946]: I1203 09:09:03.783020 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fxqkt" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerName="registry-server" containerID="cri-o://17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82" gracePeriod=2 Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.268519 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.440517 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-catalog-content\") pod \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.440607 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-utilities\") pod \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.440693 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klpqk\" (UniqueName: \"kubernetes.io/projected/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-kube-api-access-klpqk\") pod \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\" (UID: \"6e13dce2-6b6f-4ffd-850a-9a40b522bbca\") " Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.442171 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-utilities" (OuterVolumeSpecName: "utilities") pod "6e13dce2-6b6f-4ffd-850a-9a40b522bbca" (UID: "6e13dce2-6b6f-4ffd-850a-9a40b522bbca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.449359 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-kube-api-access-klpqk" (OuterVolumeSpecName: "kube-api-access-klpqk") pod "6e13dce2-6b6f-4ffd-850a-9a40b522bbca" (UID: "6e13dce2-6b6f-4ffd-850a-9a40b522bbca"). InnerVolumeSpecName "kube-api-access-klpqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.542210 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.542457 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klpqk\" (UniqueName: \"kubernetes.io/projected/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-kube-api-access-klpqk\") on node \"crc\" DevicePath \"\"" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.554348 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e13dce2-6b6f-4ffd-850a-9a40b522bbca" (UID: "6e13dce2-6b6f-4ffd-850a-9a40b522bbca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.644634 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e13dce2-6b6f-4ffd-850a-9a40b522bbca-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.803282 4946 generic.go:334] "Generic (PLEG): container finished" podID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerID="17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82" exitCode=0 Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.803379 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxqkt" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.803388 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxqkt" event={"ID":"6e13dce2-6b6f-4ffd-850a-9a40b522bbca","Type":"ContainerDied","Data":"17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82"} Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.803973 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxqkt" event={"ID":"6e13dce2-6b6f-4ffd-850a-9a40b522bbca","Type":"ContainerDied","Data":"02ff560f30b42a7f4c5bf9ff123dc28451531ed3d357018060664ff6c598190a"} Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.804013 4946 scope.go:117] "RemoveContainer" containerID="17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.842872 4946 scope.go:117] "RemoveContainer" containerID="46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.873230 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fxqkt"] Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.880552 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fxqkt"] Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.882635 4946 scope.go:117] "RemoveContainer" containerID="a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.911954 4946 scope.go:117] "RemoveContainer" containerID="17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82" Dec 03 09:09:04 crc kubenswrapper[4946]: E1203 09:09:04.912639 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82\": container with ID starting with 17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82 not found: ID does not exist" containerID="17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.912696 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82"} err="failed to get container status \"17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82\": rpc error: code = NotFound desc = could not find container \"17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82\": container with ID starting with 17863f657d6dedf5285f56c51dca5b7015b0efdb0903f9e63c0373c7fb109f82 not found: ID does not exist" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.912728 4946 scope.go:117] "RemoveContainer" containerID="46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a" Dec 03 09:09:04 crc kubenswrapper[4946]: E1203 09:09:04.913141 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a\": container with ID starting with 46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a not found: ID does not exist" containerID="46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.913184 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a"} err="failed to get container status \"46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a\": rpc error: code = NotFound desc = could not find container \"46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a\": container with ID starting with 46613ca15bd0df8f9e5c462ee7a3a7babc3ec696bb657cddaa40ac153e87874a not found: ID does not exist" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.913201 4946 scope.go:117] "RemoveContainer" containerID="a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b" Dec 03 09:09:04 crc kubenswrapper[4946]: E1203 09:09:04.913564 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b\": container with ID starting with a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b not found: ID does not exist" containerID="a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b" Dec 03 09:09:04 crc kubenswrapper[4946]: I1203 09:09:04.913634 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b"} err="failed to get container status \"a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b\": rpc error: code = NotFound desc = could not find container \"a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b\": container with ID starting with a6d385d47782e0f4873194be41e927cd99c0b0e93b25df40d3d20ec5d69a0e0b not found: ID does not exist" Dec 03 09:09:05 crc kubenswrapper[4946]: I1203 09:09:05.601128 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" path="/var/lib/kubelet/pods/6e13dce2-6b6f-4ffd-850a-9a40b522bbca/volumes" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.051451 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wfm8d"] Dec 03 09:10:00 crc kubenswrapper[4946]: E1203 09:10:00.052788 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerName="extract-utilities" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.052823 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerName="extract-utilities" Dec 03 09:10:00 crc kubenswrapper[4946]: E1203 09:10:00.052861 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerName="registry-server" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.052880 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerName="registry-server" Dec 03 09:10:00 crc kubenswrapper[4946]: E1203 09:10:00.052907 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerName="extract-content" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.052922 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerName="extract-content" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.053200 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e13dce2-6b6f-4ffd-850a-9a40b522bbca" containerName="registry-server" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.055617 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.078167 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wfm8d"] Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.118854 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-utilities\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.119017 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-catalog-content\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.119193 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv86j\" (UniqueName: \"kubernetes.io/projected/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-kube-api-access-vv86j\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.220875 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv86j\" (UniqueName: \"kubernetes.io/projected/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-kube-api-access-vv86j\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.221135 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-utilities\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.221246 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-catalog-content\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.221819 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-catalog-content\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.221819 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-utilities\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.252165 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv86j\" (UniqueName: \"kubernetes.io/projected/ecf4458a-21fd-4426-a0aa-0f8342d40c3b-kube-api-access-vv86j\") pod \"redhat-operators-wfm8d\" (UID: \"ecf4458a-21fd-4426-a0aa-0f8342d40c3b\") " pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.383886 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:00 crc kubenswrapper[4946]: I1203 09:10:00.809552 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wfm8d"] Dec 03 09:10:01 crc kubenswrapper[4946]: I1203 09:10:01.304456 4946 generic.go:334] "Generic (PLEG): container finished" podID="ecf4458a-21fd-4426-a0aa-0f8342d40c3b" containerID="9f799c75c68cd86485824a820078afac621273476d8aa56647e8ce4756f0c4e5" exitCode=0 Dec 03 09:10:01 crc kubenswrapper[4946]: I1203 09:10:01.304539 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfm8d" event={"ID":"ecf4458a-21fd-4426-a0aa-0f8342d40c3b","Type":"ContainerDied","Data":"9f799c75c68cd86485824a820078afac621273476d8aa56647e8ce4756f0c4e5"} Dec 03 09:10:01 crc kubenswrapper[4946]: I1203 09:10:01.304814 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfm8d" event={"ID":"ecf4458a-21fd-4426-a0aa-0f8342d40c3b","Type":"ContainerStarted","Data":"150bbdb16711966ce5c0cc19d506f644b5cd935189e5858e850463025579613f"} Dec 03 09:10:09 crc kubenswrapper[4946]: I1203 09:10:09.369345 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfm8d" event={"ID":"ecf4458a-21fd-4426-a0aa-0f8342d40c3b","Type":"ContainerStarted","Data":"8a73e9feb55b59ca2470e0a31caffc189b8a5530a730ba3e003a949a232a4ec3"} Dec 03 09:10:10 crc kubenswrapper[4946]: I1203 09:10:10.383556 4946 generic.go:334] "Generic (PLEG): container finished" podID="ecf4458a-21fd-4426-a0aa-0f8342d40c3b" containerID="8a73e9feb55b59ca2470e0a31caffc189b8a5530a730ba3e003a949a232a4ec3" exitCode=0 Dec 03 09:10:10 crc kubenswrapper[4946]: I1203 09:10:10.383627 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfm8d" event={"ID":"ecf4458a-21fd-4426-a0aa-0f8342d40c3b","Type":"ContainerDied","Data":"8a73e9feb55b59ca2470e0a31caffc189b8a5530a730ba3e003a949a232a4ec3"} Dec 03 09:10:11 crc kubenswrapper[4946]: I1203 09:10:11.395455 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfm8d" event={"ID":"ecf4458a-21fd-4426-a0aa-0f8342d40c3b","Type":"ContainerStarted","Data":"a2ffe895d9d037d6824c41ad2e2a9be46a241f871c12d1155695aba3be6a2245"} Dec 03 09:10:11 crc kubenswrapper[4946]: I1203 09:10:11.425892 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wfm8d" podStartSLOduration=1.966369136 podStartE2EDuration="11.425870337s" podCreationTimestamp="2025-12-03 09:10:00 +0000 UTC" firstStartedPulling="2025-12-03 09:10:01.306329653 +0000 UTC m=+8394.103019762" lastFinishedPulling="2025-12-03 09:10:10.765830854 +0000 UTC m=+8403.562520963" observedRunningTime="2025-12-03 09:10:11.419105696 +0000 UTC m=+8404.215795805" watchObservedRunningTime="2025-12-03 09:10:11.425870337 +0000 UTC m=+8404.222560446" Dec 03 09:10:20 crc kubenswrapper[4946]: I1203 09:10:20.385255 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:20 crc kubenswrapper[4946]: I1203 09:10:20.385820 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:20 crc kubenswrapper[4946]: I1203 09:10:20.434194 4946 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:20 crc kubenswrapper[4946]: I1203 09:10:20.498560 4946 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wfm8d" Dec 03 09:10:20 crc kubenswrapper[4946]: I1203 09:10:20.550902 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wfm8d"] Dec 03 09:10:20 crc kubenswrapper[4946]: I1203 09:10:20.668192 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8kn9"] Dec 03 09:10:20 crc kubenswrapper[4946]: I1203 09:10:20.668424 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g8kn9" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerName="registry-server" containerID="cri-o://6fd911556a7b0edb06aa8bc89a38946cf1361756b0a011df9881a47247e98c7c" gracePeriod=2 Dec 03 09:10:22 crc kubenswrapper[4946]: I1203 09:10:22.481287 4946 generic.go:334] "Generic (PLEG): container finished" podID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerID="6fd911556a7b0edb06aa8bc89a38946cf1361756b0a011df9881a47247e98c7c" exitCode=0 Dec 03 09:10:22 crc kubenswrapper[4946]: I1203 09:10:22.481384 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8kn9" event={"ID":"1f0bfec9-3a64-4320-96f0-f262aad05298","Type":"ContainerDied","Data":"6fd911556a7b0edb06aa8bc89a38946cf1361756b0a011df9881a47247e98c7c"} Dec 03 09:10:24 crc kubenswrapper[4946]: I1203 09:10:24.799964 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 09:10:24 crc kubenswrapper[4946]: I1203 09:10:24.998793 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4gsr\" (UniqueName: \"kubernetes.io/projected/1f0bfec9-3a64-4320-96f0-f262aad05298-kube-api-access-g4gsr\") pod \"1f0bfec9-3a64-4320-96f0-f262aad05298\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " Dec 03 09:10:24 crc kubenswrapper[4946]: I1203 09:10:24.998884 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-utilities\") pod \"1f0bfec9-3a64-4320-96f0-f262aad05298\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " Dec 03 09:10:24 crc kubenswrapper[4946]: I1203 09:10:24.999023 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-catalog-content\") pod \"1f0bfec9-3a64-4320-96f0-f262aad05298\" (UID: \"1f0bfec9-3a64-4320-96f0-f262aad05298\") " Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:24.999995 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-utilities" (OuterVolumeSpecName: "utilities") pod "1f0bfec9-3a64-4320-96f0-f262aad05298" (UID: "1f0bfec9-3a64-4320-96f0-f262aad05298"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.006049 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f0bfec9-3a64-4320-96f0-f262aad05298-kube-api-access-g4gsr" (OuterVolumeSpecName: "kube-api-access-g4gsr") pod "1f0bfec9-3a64-4320-96f0-f262aad05298" (UID: "1f0bfec9-3a64-4320-96f0-f262aad05298"). InnerVolumeSpecName "kube-api-access-g4gsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.100959 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4gsr\" (UniqueName: \"kubernetes.io/projected/1f0bfec9-3a64-4320-96f0-f262aad05298-kube-api-access-g4gsr\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.100997 4946 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.114373 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f0bfec9-3a64-4320-96f0-f262aad05298" (UID: "1f0bfec9-3a64-4320-96f0-f262aad05298"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.201916 4946 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f0bfec9-3a64-4320-96f0-f262aad05298-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.513906 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8kn9" event={"ID":"1f0bfec9-3a64-4320-96f0-f262aad05298","Type":"ContainerDied","Data":"295904b2187cd932d5b9be51d1700a2a0c0e7673ada32fb745b99dcaf81908c4"} Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.514017 4946 scope.go:117] "RemoveContainer" containerID="6fd911556a7b0edb06aa8bc89a38946cf1361756b0a011df9881a47247e98c7c" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.514037 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8kn9" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.554595 4946 scope.go:117] "RemoveContainer" containerID="a445de542ad83bc6bf7796fdc0de31f590cfeb65f5fecef23c010a04b7fa5190" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.565349 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8kn9"] Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.574115 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g8kn9"] Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.584498 4946 scope.go:117] "RemoveContainer" containerID="b69758e7a5c0c87afadbc8ccd9bb76320ebaed1353de0217c53ad2bb2ded14f6" Dec 03 09:10:25 crc kubenswrapper[4946]: I1203 09:10:25.608076 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" path="/var/lib/kubelet/pods/1f0bfec9-3a64-4320-96f0-f262aad05298/volumes" Dec 03 09:10:53 crc kubenswrapper[4946]: I1203 09:10:53.039403 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:10:53 crc kubenswrapper[4946]: I1203 09:10:53.040112 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.446132 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dlnpd/must-gather-78pd2"] Dec 03 09:11:06 crc kubenswrapper[4946]: E1203 09:11:06.447011 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerName="extract-content" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.447024 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerName="extract-content" Dec 03 09:11:06 crc kubenswrapper[4946]: E1203 09:11:06.447042 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerName="extract-utilities" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.447070 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerName="extract-utilities" Dec 03 09:11:06 crc kubenswrapper[4946]: E1203 09:11:06.447090 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerName="registry-server" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.447096 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerName="registry-server" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.447249 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f0bfec9-3a64-4320-96f0-f262aad05298" containerName="registry-server" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.447998 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.452880 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-dlnpd/must-gather-78pd2"] Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.453156 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-dlnpd"/"kube-root-ca.crt" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.454323 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-dlnpd"/"openshift-service-ca.crt" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.454341 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-dlnpd"/"default-dockercfg-p8dhd" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.547926 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8306927b-521d-4ebd-abc8-9eba912684ca-must-gather-output\") pod \"must-gather-78pd2\" (UID: \"8306927b-521d-4ebd-abc8-9eba912684ca\") " pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.547991 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z7jt\" (UniqueName: \"kubernetes.io/projected/8306927b-521d-4ebd-abc8-9eba912684ca-kube-api-access-2z7jt\") pod \"must-gather-78pd2\" (UID: \"8306927b-521d-4ebd-abc8-9eba912684ca\") " pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.649306 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z7jt\" (UniqueName: \"kubernetes.io/projected/8306927b-521d-4ebd-abc8-9eba912684ca-kube-api-access-2z7jt\") pod \"must-gather-78pd2\" (UID: \"8306927b-521d-4ebd-abc8-9eba912684ca\") " pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.649455 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8306927b-521d-4ebd-abc8-9eba912684ca-must-gather-output\") pod \"must-gather-78pd2\" (UID: \"8306927b-521d-4ebd-abc8-9eba912684ca\") " pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.649958 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8306927b-521d-4ebd-abc8-9eba912684ca-must-gather-output\") pod \"must-gather-78pd2\" (UID: \"8306927b-521d-4ebd-abc8-9eba912684ca\") " pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.690088 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z7jt\" (UniqueName: \"kubernetes.io/projected/8306927b-521d-4ebd-abc8-9eba912684ca-kube-api-access-2z7jt\") pod \"must-gather-78pd2\" (UID: \"8306927b-521d-4ebd-abc8-9eba912684ca\") " pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:11:06 crc kubenswrapper[4946]: I1203 09:11:06.764348 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:11:07 crc kubenswrapper[4946]: I1203 09:11:07.068318 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-dlnpd/must-gather-78pd2"] Dec 03 09:11:07 crc kubenswrapper[4946]: I1203 09:11:07.875421 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dlnpd/must-gather-78pd2" event={"ID":"8306927b-521d-4ebd-abc8-9eba912684ca","Type":"ContainerStarted","Data":"b22962150f08c89693af793bdc5bd6c993680b3be7a65dc277ceaf9fd9ab03ba"} Dec 03 09:11:11 crc kubenswrapper[4946]: I1203 09:11:11.923279 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dlnpd/must-gather-78pd2" event={"ID":"8306927b-521d-4ebd-abc8-9eba912684ca","Type":"ContainerStarted","Data":"663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58"} Dec 03 09:11:11 crc kubenswrapper[4946]: I1203 09:11:11.924080 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dlnpd/must-gather-78pd2" event={"ID":"8306927b-521d-4ebd-abc8-9eba912684ca","Type":"ContainerStarted","Data":"63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f"} Dec 03 09:11:11 crc kubenswrapper[4946]: I1203 09:11:11.945028 4946 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-dlnpd/must-gather-78pd2" podStartSLOduration=2.318848786 podStartE2EDuration="5.945004719s" podCreationTimestamp="2025-12-03 09:11:06 +0000 UTC" firstStartedPulling="2025-12-03 09:11:07.073961782 +0000 UTC m=+8459.870651891" lastFinishedPulling="2025-12-03 09:11:10.700117725 +0000 UTC m=+8463.496807824" observedRunningTime="2025-12-03 09:11:11.942689468 +0000 UTC m=+8464.739379617" watchObservedRunningTime="2025-12-03 09:11:11.945004719 +0000 UTC m=+8464.741694848" Dec 03 09:11:23 crc kubenswrapper[4946]: I1203 09:11:23.039357 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:11:23 crc kubenswrapper[4946]: I1203 09:11:23.040105 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.039605 4946 patch_prober.go:28] interesting pod/machine-config-daemon-6bt2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.041522 4946 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.041695 4946 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.042451 4946 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0"} pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.042621 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerName="machine-config-daemon" containerID="cri-o://097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" gracePeriod=600 Dec 03 09:11:53 crc kubenswrapper[4946]: E1203 09:11:53.190474 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.255094 4946 generic.go:334] "Generic (PLEG): container finished" podID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" exitCode=0 Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.255202 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerDied","Data":"097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0"} Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.255649 4946 scope.go:117] "RemoveContainer" containerID="8a3899d0e6f82103ad9642f8af450b7d00441adf16c0a905df352bdafc37da27" Dec 03 09:11:53 crc kubenswrapper[4946]: I1203 09:11:53.256874 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:11:53 crc kubenswrapper[4946]: E1203 09:11:53.257575 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:12:04 crc kubenswrapper[4946]: I1203 09:12:04.592504 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:12:04 crc kubenswrapper[4946]: E1203 09:12:04.594139 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:12:08 crc kubenswrapper[4946]: I1203 09:12:08.447451 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p_e21599fd-a0aa-46eb-80d9-a50c426bb629/util/0.log" Dec 03 09:12:08 crc kubenswrapper[4946]: I1203 09:12:08.673721 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p_e21599fd-a0aa-46eb-80d9-a50c426bb629/pull/0.log" Dec 03 09:12:08 crc kubenswrapper[4946]: I1203 09:12:08.691101 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p_e21599fd-a0aa-46eb-80d9-a50c426bb629/util/0.log" Dec 03 09:12:08 crc kubenswrapper[4946]: I1203 09:12:08.709486 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p_e21599fd-a0aa-46eb-80d9-a50c426bb629/pull/0.log" Dec 03 09:12:08 crc kubenswrapper[4946]: I1203 09:12:08.935318 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p_e21599fd-a0aa-46eb-80d9-a50c426bb629/pull/0.log" Dec 03 09:12:08 crc kubenswrapper[4946]: I1203 09:12:08.938227 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p_e21599fd-a0aa-46eb-80d9-a50c426bb629/util/0.log" Dec 03 09:12:08 crc kubenswrapper[4946]: I1203 09:12:08.978286 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_98dc3bd0b5c63de8bc52e3558b9d3e72fafafb6fd127fd2510d22068644xb6p_e21599fd-a0aa-46eb-80d9-a50c426bb629/extract/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.110290 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-v62hf_881fe69d-7e12-4250-8589-76096d3318d9/kube-rbac-proxy/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.151292 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-v62hf_881fe69d-7e12-4250-8589-76096d3318d9/manager/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.233473 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-v7djg_592abc1b-947b-4b47-965f-0110fa5e74e3/kube-rbac-proxy/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.285433 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-v7djg_592abc1b-947b-4b47-965f-0110fa5e74e3/manager/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.340855 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-fwbh4_feaece87-4f9e-4945-8faa-c49487a57c2f/kube-rbac-proxy/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.412885 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-fwbh4_feaece87-4f9e-4945-8faa-c49487a57c2f/manager/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.502791 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-rkv2m_f050d6c8-74b6-42cb-943b-6e372c3e0dfe/kube-rbac-proxy/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.636840 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-7cwn9_afb5f2e7-fb0a-425b-9a3c-cff10fdec757/kube-rbac-proxy/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.656526 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-rkv2m_f050d6c8-74b6-42cb-943b-6e372c3e0dfe/manager/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.718912 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-7cwn9_afb5f2e7-fb0a-425b-9a3c-cff10fdec757/manager/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.863282 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-sv727_bd12d71a-063b-445c-876b-aab60ef74f35/manager/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.868120 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-sv727_bd12d71a-063b-445c-876b-aab60ef74f35/kube-rbac-proxy/0.log" Dec 03 09:12:09 crc kubenswrapper[4946]: I1203 09:12:09.986187 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-lvk95_26f08bfc-aab0-4d18-a223-8e93309c15b0/kube-rbac-proxy/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.100399 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-q88nf_29fe18fb-39d7-4003-8171-0e47aa1d1b18/kube-rbac-proxy/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.163829 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-q88nf_29fe18fb-39d7-4003-8171-0e47aa1d1b18/manager/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.164145 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-lvk95_26f08bfc-aab0-4d18-a223-8e93309c15b0/manager/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.284325 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-b68fp_f55d3e9f-d953-4672-9438-4f8db6a0af1b/kube-rbac-proxy/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.351057 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-b68fp_f55d3e9f-d953-4672-9438-4f8db6a0af1b/manager/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.493663 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-q27df_9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81/kube-rbac-proxy/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.539809 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-q27df_9fe6ceb6-877f-4d5e-9c47-cb089ebdfe81/manager/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.627751 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2hnpw_4110947d-509f-4bb7-bb92-e52ebb362340/kube-rbac-proxy/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.724467 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2hnpw_4110947d-509f-4bb7-bb92-e52ebb362340/manager/0.log" Dec 03 09:12:10 crc kubenswrapper[4946]: I1203 09:12:10.970734 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-9n79w_f6febc9c-4437-4f94-8996-f4ece33a42f2/kube-rbac-proxy/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.015867 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-9n79w_f6febc9c-4437-4f94-8996-f4ece33a42f2/manager/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.127652 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-g6cf2_59735e89-ef8b-4887-9af2-d2bd5d3f7bb0/kube-rbac-proxy/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.248653 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-g6cf2_59735e89-ef8b-4887-9af2-d2bd5d3f7bb0/manager/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.304883 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-xlsd9_77e8b06b-4c6d-478a-ac26-230010446011/kube-rbac-proxy/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.330176 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-xlsd9_77e8b06b-4c6d-478a-ac26-230010446011/manager/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.440123 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55d86b6686k9mhl_ee6d6741-bc1e-4ed2-a54c-608c37d5da97/kube-rbac-proxy/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.494306 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55d86b6686k9mhl_ee6d6741-bc1e-4ed2-a54c-608c37d5da97/manager/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.903892 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7dd5c7bb7c-hh9qp_9cb25a27-f4e6-4a78-a9f3-c8527cbc7b77/operator/0.log" Dec 03 09:12:11 crc kubenswrapper[4946]: I1203 09:12:11.981069 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fq7s4_1b592596-16d1-410d-953b-ad7c04d2f6a4/registry-server/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.025320 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-zws5c_06dff8f3-1b1f-4905-8976-ad14fd8b0b3f/kube-rbac-proxy/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.243582 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-p28rk_4bbd7e74-3c10-47ea-ac92-2845c974a128/kube-rbac-proxy/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.260939 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-p28rk_4bbd7e74-3c10-47ea-ac92-2845c974a128/manager/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.273669 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-zws5c_06dff8f3-1b1f-4905-8976-ad14fd8b0b3f/manager/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.410575 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-9f56fc979-wtn4r_7c469591-852f-4294-bd9e-8d1dc58e5d75/manager/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.443476 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-z2k4w_e20ea5bc-56c0-49ac-a27c-e6e93778a09a/operator/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.469048 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-w9prn_ab0a2041-fd3d-49dc-b184-fdd603df844c/kube-rbac-proxy/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.566634 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-w9prn_ab0a2041-fd3d-49dc-b184-fdd603df844c/manager/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.648191 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nq7sd_a063e091-3cc0-4966-a2d7-69bcba9885f3/kube-rbac-proxy/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.716619 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nq7sd_a063e091-3cc0-4966-a2d7-69bcba9885f3/manager/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.755903 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-l4hnh_12f0ad36-d30d-4d23-8677-8d452c5ea978/kube-rbac-proxy/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.822170 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-l4hnh_12f0ad36-d30d-4d23-8677-8d452c5ea978/manager/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.901118 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-5p4zw_3efd94f9-2bf7-404c-ae24-6884ead0de15/kube-rbac-proxy/0.log" Dec 03 09:12:12 crc kubenswrapper[4946]: I1203 09:12:12.959270 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-5p4zw_3efd94f9-2bf7-404c-ae24-6884ead0de15/manager/0.log" Dec 03 09:12:15 crc kubenswrapper[4946]: I1203 09:12:15.592837 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:12:15 crc kubenswrapper[4946]: E1203 09:12:15.593393 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:12:26 crc kubenswrapper[4946]: I1203 09:12:26.593841 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:12:26 crc kubenswrapper[4946]: E1203 09:12:26.594640 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:12:32 crc kubenswrapper[4946]: I1203 09:12:32.180656 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-b5trw_0acb31c7-623c-40b7-889c-4a9a45b55cf1/control-plane-machine-set-operator/0.log" Dec 03 09:12:32 crc kubenswrapper[4946]: I1203 09:12:32.354429 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vcrrg_6ff611ff-6a55-435a-9723-df1b38741952/kube-rbac-proxy/0.log" Dec 03 09:12:32 crc kubenswrapper[4946]: I1203 09:12:32.405698 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vcrrg_6ff611ff-6a55-435a-9723-df1b38741952/machine-api-operator/0.log" Dec 03 09:12:41 crc kubenswrapper[4946]: I1203 09:12:41.592343 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:12:41 crc kubenswrapper[4946]: E1203 09:12:41.593061 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:12:45 crc kubenswrapper[4946]: I1203 09:12:45.046116 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-xcngx_babf0398-4ef7-4498-be2c-b970cc3db40c/cert-manager-controller/0.log" Dec 03 09:12:45 crc kubenswrapper[4946]: I1203 09:12:45.176590 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-dcppn_07b6bc6f-2572-4b9b-99f2-c4ba512dcede/cert-manager-cainjector/0.log" Dec 03 09:12:45 crc kubenswrapper[4946]: I1203 09:12:45.240524 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-wplrd_1837c927-ee2e-4452-aec6-9306ab8dff65/cert-manager-webhook/0.log" Dec 03 09:12:52 crc kubenswrapper[4946]: I1203 09:12:52.593138 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:12:52 crc kubenswrapper[4946]: E1203 09:12:52.594071 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:12:57 crc kubenswrapper[4946]: I1203 09:12:57.744801 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5jbpq_ae0ab229-8baf-4c5b-a452-2fdd714946ea/nmstate-console-plugin/0.log" Dec 03 09:12:57 crc kubenswrapper[4946]: I1203 09:12:57.902155 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vmz8j_7219ceec-2cc3-4663-bf50-11ef855ec612/nmstate-handler/0.log" Dec 03 09:12:57 crc kubenswrapper[4946]: I1203 09:12:57.998418 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-smzvj_1e225c74-69b1-4094-9310-7e790d712264/kube-rbac-proxy/0.log" Dec 03 09:12:58 crc kubenswrapper[4946]: I1203 09:12:58.000100 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-smzvj_1e225c74-69b1-4094-9310-7e790d712264/nmstate-metrics/0.log" Dec 03 09:12:58 crc kubenswrapper[4946]: I1203 09:12:58.183803 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-xjqdt_0bcd8282-02a8-4389-bc9c-bb6d2b0904df/nmstate-operator/0.log" Dec 03 09:12:58 crc kubenswrapper[4946]: I1203 09:12:58.188247 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-jx9l9_f636b80e-63df-4797-9720-bc7daa273204/nmstate-webhook/0.log" Dec 03 09:13:05 crc kubenswrapper[4946]: I1203 09:13:05.592938 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:13:05 crc kubenswrapper[4946]: E1203 09:13:05.593717 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:13:13 crc kubenswrapper[4946]: I1203 09:13:13.677938 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-5bh6f_3ff6d4b8-37a3-43af-bac0-00a51011a791/kube-rbac-proxy/0.log" Dec 03 09:13:13 crc kubenswrapper[4946]: I1203 09:13:13.893030 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-5bh6f_3ff6d4b8-37a3-43af-bac0-00a51011a791/controller/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.046130 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-frr-files/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.189633 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-reloader/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.233910 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-metrics/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.261260 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-frr-files/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.290177 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-reloader/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.433883 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-frr-files/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.440441 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-reloader/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.489971 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-metrics/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.519216 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-metrics/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.635027 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-frr-files/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.686292 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-metrics/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.693229 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/cp-reloader/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.730583 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/controller/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.862630 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/frr-metrics/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.893090 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/kube-rbac-proxy/0.log" Dec 03 09:13:14 crc kubenswrapper[4946]: I1203 09:13:14.930985 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/kube-rbac-proxy-frr/0.log" Dec 03 09:13:15 crc kubenswrapper[4946]: I1203 09:13:15.090082 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/reloader/0.log" Dec 03 09:13:15 crc kubenswrapper[4946]: I1203 09:13:15.294389 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-kb5zx_6c6d5045-fb6d-407d-b044-f42503aa1ce1/frr-k8s-webhook-server/0.log" Dec 03 09:13:15 crc kubenswrapper[4946]: I1203 09:13:15.400833 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5597d58577-85fkn_9cefae38-3773-4b3c-b0a7-fd4c4c3fea14/manager/0.log" Dec 03 09:13:15 crc kubenswrapper[4946]: I1203 09:13:15.530814 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-78b68b94b4-klptw_57c3ab00-3c9b-41d1-8c7e-9c51b8fd5b3d/webhook-server/0.log" Dec 03 09:13:15 crc kubenswrapper[4946]: I1203 09:13:15.678438 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fsnxs_55fde23e-25e3-4aca-88a3-878ae3117972/kube-rbac-proxy/0.log" Dec 03 09:13:16 crc kubenswrapper[4946]: I1203 09:13:16.166885 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fsnxs_55fde23e-25e3-4aca-88a3-878ae3117972/speaker/0.log" Dec 03 09:13:16 crc kubenswrapper[4946]: I1203 09:13:16.209682 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4h989_34f70191-7d9a-4c12-8a82-f2927ac34e75/frr/0.log" Dec 03 09:13:20 crc kubenswrapper[4946]: I1203 09:13:20.593550 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:13:20 crc kubenswrapper[4946]: E1203 09:13:20.595676 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:13:30 crc kubenswrapper[4946]: I1203 09:13:30.424412 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj_268ff739-d59c-4d7c-a413-bcb696ef8c86/util/0.log" Dec 03 09:13:30 crc kubenswrapper[4946]: I1203 09:13:30.570254 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj_268ff739-d59c-4d7c-a413-bcb696ef8c86/util/0.log" Dec 03 09:13:30 crc kubenswrapper[4946]: I1203 09:13:30.602767 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj_268ff739-d59c-4d7c-a413-bcb696ef8c86/pull/0.log" Dec 03 09:13:30 crc kubenswrapper[4946]: I1203 09:13:30.667044 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj_268ff739-d59c-4d7c-a413-bcb696ef8c86/pull/0.log" Dec 03 09:13:30 crc kubenswrapper[4946]: I1203 09:13:30.777366 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj_268ff739-d59c-4d7c-a413-bcb696ef8c86/util/0.log" Dec 03 09:13:30 crc kubenswrapper[4946]: I1203 09:13:30.841342 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj_268ff739-d59c-4d7c-a413-bcb696ef8c86/pull/0.log" Dec 03 09:13:30 crc kubenswrapper[4946]: I1203 09:13:30.845623 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4kkdj_268ff739-d59c-4d7c-a413-bcb696ef8c86/extract/0.log" Dec 03 09:13:30 crc kubenswrapper[4946]: I1203 09:13:30.950416 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2_1e74163b-85c4-4470-b899-a40817bcfae9/util/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.177002 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2_1e74163b-85c4-4470-b899-a40817bcfae9/pull/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.185022 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2_1e74163b-85c4-4470-b899-a40817bcfae9/util/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.191157 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2_1e74163b-85c4-4470-b899-a40817bcfae9/pull/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.418782 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2_1e74163b-85c4-4470-b899-a40817bcfae9/util/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.433588 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2_1e74163b-85c4-4470-b899-a40817bcfae9/pull/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.466595 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f2f6k2_1e74163b-85c4-4470-b899-a40817bcfae9/extract/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.605877 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq_503cf5b8-948c-468a-a0d2-d17c0b2e7fbf/util/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.764967 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq_503cf5b8-948c-468a-a0d2-d17c0b2e7fbf/util/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.773259 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq_503cf5b8-948c-468a-a0d2-d17c0b2e7fbf/pull/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.782137 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq_503cf5b8-948c-468a-a0d2-d17c0b2e7fbf/pull/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.928527 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq_503cf5b8-948c-468a-a0d2-d17c0b2e7fbf/util/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.929682 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq_503cf5b8-948c-468a-a0d2-d17c0b2e7fbf/extract/0.log" Dec 03 09:13:31 crc kubenswrapper[4946]: I1203 09:13:31.937296 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c5kdq_503cf5b8-948c-468a-a0d2-d17c0b2e7fbf/pull/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.056843 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hsdv2_e5e5669e-63e7-4bae-892a-04d994f9b400/extract-utilities/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.227508 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hsdv2_e5e5669e-63e7-4bae-892a-04d994f9b400/extract-utilities/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.288116 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hsdv2_e5e5669e-63e7-4bae-892a-04d994f9b400/extract-content/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.293263 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hsdv2_e5e5669e-63e7-4bae-892a-04d994f9b400/extract-content/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.455209 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hsdv2_e5e5669e-63e7-4bae-892a-04d994f9b400/extract-utilities/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.461383 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hsdv2_e5e5669e-63e7-4bae-892a-04d994f9b400/extract-content/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.646374 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gtn8p_6ca0020d-490e-4efe-a145-a4bee0006e53/extract-utilities/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.894870 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gtn8p_6ca0020d-490e-4efe-a145-a4bee0006e53/extract-content/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.910368 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gtn8p_6ca0020d-490e-4efe-a145-a4bee0006e53/extract-content/0.log" Dec 03 09:13:32 crc kubenswrapper[4946]: I1203 09:13:32.939867 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gtn8p_6ca0020d-490e-4efe-a145-a4bee0006e53/extract-utilities/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.082854 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gtn8p_6ca0020d-490e-4efe-a145-a4bee0006e53/extract-content/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.107888 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hsdv2_e5e5669e-63e7-4bae-892a-04d994f9b400/registry-server/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.128875 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gtn8p_6ca0020d-490e-4efe-a145-a4bee0006e53/extract-utilities/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.336774 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-g2gmk_80417841-c6f8-4069-99bf-c98a62caa7c9/marketplace-operator/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.522429 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m2bgp_90735cf4-f312-4185-adf9-eb2112863a50/extract-utilities/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.734917 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m2bgp_90735cf4-f312-4185-adf9-eb2112863a50/extract-content/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.795577 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m2bgp_90735cf4-f312-4185-adf9-eb2112863a50/extract-content/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.828821 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m2bgp_90735cf4-f312-4185-adf9-eb2112863a50/extract-utilities/0.log" Dec 03 09:13:33 crc kubenswrapper[4946]: I1203 09:13:33.977084 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m2bgp_90735cf4-f312-4185-adf9-eb2112863a50/extract-utilities/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.014059 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m2bgp_90735cf4-f312-4185-adf9-eb2112863a50/extract-content/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.135096 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gtn8p_6ca0020d-490e-4efe-a145-a4bee0006e53/registry-server/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.181176 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wfm8d_ecf4458a-21fd-4426-a0aa-0f8342d40c3b/extract-utilities/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.372795 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m2bgp_90735cf4-f312-4185-adf9-eb2112863a50/registry-server/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.380591 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wfm8d_ecf4458a-21fd-4426-a0aa-0f8342d40c3b/extract-content/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.386792 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wfm8d_ecf4458a-21fd-4426-a0aa-0f8342d40c3b/extract-utilities/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.389866 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wfm8d_ecf4458a-21fd-4426-a0aa-0f8342d40c3b/extract-content/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.538006 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wfm8d_ecf4458a-21fd-4426-a0aa-0f8342d40c3b/extract-utilities/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.551345 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wfm8d_ecf4458a-21fd-4426-a0aa-0f8342d40c3b/extract-content/0.log" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.592208 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:13:34 crc kubenswrapper[4946]: E1203 09:13:34.592548 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:13:34 crc kubenswrapper[4946]: I1203 09:13:34.654442 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wfm8d_ecf4458a-21fd-4426-a0aa-0f8342d40c3b/registry-server/0.log" Dec 03 09:13:47 crc kubenswrapper[4946]: I1203 09:13:47.596628 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:13:47 crc kubenswrapper[4946]: E1203 09:13:47.597539 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:13:58 crc kubenswrapper[4946]: I1203 09:13:58.593229 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:13:58 crc kubenswrapper[4946]: E1203 09:13:58.594013 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:14:12 crc kubenswrapper[4946]: I1203 09:14:12.593451 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:14:12 crc kubenswrapper[4946]: E1203 09:14:12.594816 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:14:23 crc kubenswrapper[4946]: I1203 09:14:23.593565 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:14:23 crc kubenswrapper[4946]: E1203 09:14:23.594631 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:14:36 crc kubenswrapper[4946]: I1203 09:14:36.593524 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:14:36 crc kubenswrapper[4946]: E1203 09:14:36.595123 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:14:40 crc kubenswrapper[4946]: I1203 09:14:40.290512 4946 generic.go:334] "Generic (PLEG): container finished" podID="8306927b-521d-4ebd-abc8-9eba912684ca" containerID="63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f" exitCode=0 Dec 03 09:14:40 crc kubenswrapper[4946]: I1203 09:14:40.290666 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dlnpd/must-gather-78pd2" event={"ID":"8306927b-521d-4ebd-abc8-9eba912684ca","Type":"ContainerDied","Data":"63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f"} Dec 03 09:14:40 crc kubenswrapper[4946]: I1203 09:14:40.291456 4946 scope.go:117] "RemoveContainer" containerID="63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f" Dec 03 09:14:41 crc kubenswrapper[4946]: I1203 09:14:41.060155 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dlnpd_must-gather-78pd2_8306927b-521d-4ebd-abc8-9eba912684ca/gather/0.log" Dec 03 09:14:48 crc kubenswrapper[4946]: I1203 09:14:48.572317 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dlnpd/must-gather-78pd2"] Dec 03 09:14:48 crc kubenswrapper[4946]: I1203 09:14:48.572868 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dlnpd/must-gather-78pd2"] Dec 03 09:14:48 crc kubenswrapper[4946]: I1203 09:14:48.573072 4946 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-dlnpd/must-gather-78pd2" podUID="8306927b-521d-4ebd-abc8-9eba912684ca" containerName="copy" containerID="cri-o://663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58" gracePeriod=2 Dec 03 09:14:48 crc kubenswrapper[4946]: I1203 09:14:48.956185 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dlnpd_must-gather-78pd2_8306927b-521d-4ebd-abc8-9eba912684ca/copy/0.log" Dec 03 09:14:48 crc kubenswrapper[4946]: I1203 09:14:48.957161 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.113495 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z7jt\" (UniqueName: \"kubernetes.io/projected/8306927b-521d-4ebd-abc8-9eba912684ca-kube-api-access-2z7jt\") pod \"8306927b-521d-4ebd-abc8-9eba912684ca\" (UID: \"8306927b-521d-4ebd-abc8-9eba912684ca\") " Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.113880 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8306927b-521d-4ebd-abc8-9eba912684ca-must-gather-output\") pod \"8306927b-521d-4ebd-abc8-9eba912684ca\" (UID: \"8306927b-521d-4ebd-abc8-9eba912684ca\") " Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.123143 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8306927b-521d-4ebd-abc8-9eba912684ca-kube-api-access-2z7jt" (OuterVolumeSpecName: "kube-api-access-2z7jt") pod "8306927b-521d-4ebd-abc8-9eba912684ca" (UID: "8306927b-521d-4ebd-abc8-9eba912684ca"). InnerVolumeSpecName "kube-api-access-2z7jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.215305 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z7jt\" (UniqueName: \"kubernetes.io/projected/8306927b-521d-4ebd-abc8-9eba912684ca-kube-api-access-2z7jt\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.225579 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8306927b-521d-4ebd-abc8-9eba912684ca-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "8306927b-521d-4ebd-abc8-9eba912684ca" (UID: "8306927b-521d-4ebd-abc8-9eba912684ca"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.317972 4946 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8306927b-521d-4ebd-abc8-9eba912684ca-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.385464 4946 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dlnpd_must-gather-78pd2_8306927b-521d-4ebd-abc8-9eba912684ca/copy/0.log" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.385957 4946 generic.go:334] "Generic (PLEG): container finished" podID="8306927b-521d-4ebd-abc8-9eba912684ca" containerID="663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58" exitCode=143 Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.386011 4946 scope.go:117] "RemoveContainer" containerID="663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.386101 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dlnpd/must-gather-78pd2" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.405458 4946 scope.go:117] "RemoveContainer" containerID="63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.458794 4946 scope.go:117] "RemoveContainer" containerID="663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58" Dec 03 09:14:49 crc kubenswrapper[4946]: E1203 09:14:49.459456 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58\": container with ID starting with 663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58 not found: ID does not exist" containerID="663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.459516 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58"} err="failed to get container status \"663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58\": rpc error: code = NotFound desc = could not find container \"663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58\": container with ID starting with 663752eeed075f0b66fbba3ade32c0daceb45c99849ba3bb251775fc2ce84b58 not found: ID does not exist" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.459538 4946 scope.go:117] "RemoveContainer" containerID="63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f" Dec 03 09:14:49 crc kubenswrapper[4946]: E1203 09:14:49.459776 4946 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f\": container with ID starting with 63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f not found: ID does not exist" containerID="63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.459802 4946 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f"} err="failed to get container status \"63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f\": rpc error: code = NotFound desc = could not find container \"63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f\": container with ID starting with 63059f1b8fcae5bcf8751a97bd4d4e35d70ddebcc024c27d06ca766609137a1f not found: ID does not exist" Dec 03 09:14:49 crc kubenswrapper[4946]: I1203 09:14:49.601590 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8306927b-521d-4ebd-abc8-9eba912684ca" path="/var/lib/kubelet/pods/8306927b-521d-4ebd-abc8-9eba912684ca/volumes" Dec 03 09:14:51 crc kubenswrapper[4946]: I1203 09:14:51.593641 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:14:51 crc kubenswrapper[4946]: E1203 09:14:51.594425 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.194723 4946 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95"] Dec 03 09:15:00 crc kubenswrapper[4946]: E1203 09:15:00.195930 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8306927b-521d-4ebd-abc8-9eba912684ca" containerName="gather" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.195954 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8306927b-521d-4ebd-abc8-9eba912684ca" containerName="gather" Dec 03 09:15:00 crc kubenswrapper[4946]: E1203 09:15:00.195999 4946 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8306927b-521d-4ebd-abc8-9eba912684ca" containerName="copy" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.196012 4946 state_mem.go:107] "Deleted CPUSet assignment" podUID="8306927b-521d-4ebd-abc8-9eba912684ca" containerName="copy" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.196279 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8306927b-521d-4ebd-abc8-9eba912684ca" containerName="gather" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.196304 4946 memory_manager.go:354] "RemoveStaleState removing state" podUID="8306927b-521d-4ebd-abc8-9eba912684ca" containerName="copy" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.197273 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.202182 4946 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.202768 4946 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.212954 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95"] Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.385369 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2cc2a458-5dcd-413e-b13d-90ec904a7d40-config-volume\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.385436 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctk22\" (UniqueName: \"kubernetes.io/projected/2cc2a458-5dcd-413e-b13d-90ec904a7d40-kube-api-access-ctk22\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.385473 4946 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2cc2a458-5dcd-413e-b13d-90ec904a7d40-secret-volume\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.487677 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2cc2a458-5dcd-413e-b13d-90ec904a7d40-config-volume\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.488297 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctk22\" (UniqueName: \"kubernetes.io/projected/2cc2a458-5dcd-413e-b13d-90ec904a7d40-kube-api-access-ctk22\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.488652 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2cc2a458-5dcd-413e-b13d-90ec904a7d40-config-volume\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.488727 4946 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2cc2a458-5dcd-413e-b13d-90ec904a7d40-secret-volume\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.497985 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2cc2a458-5dcd-413e-b13d-90ec904a7d40-secret-volume\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.509586 4946 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctk22\" (UniqueName: \"kubernetes.io/projected/2cc2a458-5dcd-413e-b13d-90ec904a7d40-kube-api-access-ctk22\") pod \"collect-profiles-29412555-xgm95\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.534522 4946 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:00 crc kubenswrapper[4946]: I1203 09:15:00.835527 4946 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95"] Dec 03 09:15:01 crc kubenswrapper[4946]: I1203 09:15:01.486527 4946 generic.go:334] "Generic (PLEG): container finished" podID="2cc2a458-5dcd-413e-b13d-90ec904a7d40" containerID="0f375ff98b07ed010664b8efbd0b8de2dd87273259d3704a548cb80d5edc3a8a" exitCode=0 Dec 03 09:15:01 crc kubenswrapper[4946]: I1203 09:15:01.486813 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" event={"ID":"2cc2a458-5dcd-413e-b13d-90ec904a7d40","Type":"ContainerDied","Data":"0f375ff98b07ed010664b8efbd0b8de2dd87273259d3704a548cb80d5edc3a8a"} Dec 03 09:15:01 crc kubenswrapper[4946]: I1203 09:15:01.486840 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" event={"ID":"2cc2a458-5dcd-413e-b13d-90ec904a7d40","Type":"ContainerStarted","Data":"5dce3c036e4f3d926f177885f9d1a48af8ff60611a97d5c0f301c86371b50207"} Dec 03 09:15:02 crc kubenswrapper[4946]: I1203 09:15:02.920870 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.028514 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2cc2a458-5dcd-413e-b13d-90ec904a7d40-secret-volume\") pod \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.028655 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2cc2a458-5dcd-413e-b13d-90ec904a7d40-config-volume\") pod \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.028685 4946 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctk22\" (UniqueName: \"kubernetes.io/projected/2cc2a458-5dcd-413e-b13d-90ec904a7d40-kube-api-access-ctk22\") pod \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\" (UID: \"2cc2a458-5dcd-413e-b13d-90ec904a7d40\") " Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.029625 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cc2a458-5dcd-413e-b13d-90ec904a7d40-config-volume" (OuterVolumeSpecName: "config-volume") pod "2cc2a458-5dcd-413e-b13d-90ec904a7d40" (UID: "2cc2a458-5dcd-413e-b13d-90ec904a7d40"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.033665 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cc2a458-5dcd-413e-b13d-90ec904a7d40-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2cc2a458-5dcd-413e-b13d-90ec904a7d40" (UID: "2cc2a458-5dcd-413e-b13d-90ec904a7d40"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.034441 4946 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cc2a458-5dcd-413e-b13d-90ec904a7d40-kube-api-access-ctk22" (OuterVolumeSpecName: "kube-api-access-ctk22") pod "2cc2a458-5dcd-413e-b13d-90ec904a7d40" (UID: "2cc2a458-5dcd-413e-b13d-90ec904a7d40"). InnerVolumeSpecName "kube-api-access-ctk22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.130726 4946 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2cc2a458-5dcd-413e-b13d-90ec904a7d40-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.130781 4946 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2cc2a458-5dcd-413e-b13d-90ec904a7d40-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.130794 4946 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctk22\" (UniqueName: \"kubernetes.io/projected/2cc2a458-5dcd-413e-b13d-90ec904a7d40-kube-api-access-ctk22\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.511040 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" event={"ID":"2cc2a458-5dcd-413e-b13d-90ec904a7d40","Type":"ContainerDied","Data":"5dce3c036e4f3d926f177885f9d1a48af8ff60611a97d5c0f301c86371b50207"} Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.511082 4946 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dce3c036e4f3d926f177885f9d1a48af8ff60611a97d5c0f301c86371b50207" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.511146 4946 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-xgm95" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.593726 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:15:03 crc kubenswrapper[4946]: E1203 09:15:03.594188 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:15:03 crc kubenswrapper[4946]: I1203 09:15:03.987750 4946 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m"] Dec 03 09:15:04 crc kubenswrapper[4946]: I1203 09:15:04.000165 4946 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-5jn5m"] Dec 03 09:15:05 crc kubenswrapper[4946]: I1203 09:15:05.609492 4946 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa8e01b-aee0-426e-bca2-9a4678b0e9f3" path="/var/lib/kubelet/pods/5fa8e01b-aee0-426e-bca2-9a4678b0e9f3/volumes" Dec 03 09:15:14 crc kubenswrapper[4946]: I1203 09:15:14.593421 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:15:14 crc kubenswrapper[4946]: E1203 09:15:14.594177 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:15:15 crc kubenswrapper[4946]: I1203 09:15:15.951336 4946 scope.go:117] "RemoveContainer" containerID="3a89af099c1e3524ebd038ec080c79c4f44a8fc51feeec17e9b63f664aa5795f" Dec 03 09:15:27 crc kubenswrapper[4946]: I1203 09:15:27.601477 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:15:27 crc kubenswrapper[4946]: E1203 09:15:27.602468 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:15:41 crc kubenswrapper[4946]: I1203 09:15:41.593174 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:15:41 crc kubenswrapper[4946]: E1203 09:15:41.594270 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:15:53 crc kubenswrapper[4946]: I1203 09:15:53.593243 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:15:53 crc kubenswrapper[4946]: E1203 09:15:53.594404 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:16:08 crc kubenswrapper[4946]: I1203 09:16:08.595438 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:16:08 crc kubenswrapper[4946]: E1203 09:16:08.597198 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:16:21 crc kubenswrapper[4946]: I1203 09:16:21.592717 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:16:21 crc kubenswrapper[4946]: E1203 09:16:21.593477 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:16:32 crc kubenswrapper[4946]: I1203 09:16:32.592808 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:16:32 crc kubenswrapper[4946]: E1203 09:16:32.593907 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:16:47 crc kubenswrapper[4946]: I1203 09:16:47.598568 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:16:47 crc kubenswrapper[4946]: E1203 09:16:47.599421 4946 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6bt2d_openshift-machine-config-operator(4003d158-6bdd-45bd-a68c-ca52bd7264c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" podUID="4003d158-6bdd-45bd-a68c-ca52bd7264c5" Dec 03 09:17:02 crc kubenswrapper[4946]: I1203 09:17:02.594047 4946 scope.go:117] "RemoveContainer" containerID="097d5a7b7898472e6e0ac02a95ac1d106f27be23b53ffcc3e0f0dfb3b53df8d0" Dec 03 09:17:02 crc kubenswrapper[4946]: I1203 09:17:02.893570 4946 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6bt2d" event={"ID":"4003d158-6bdd-45bd-a68c-ca52bd7264c5","Type":"ContainerStarted","Data":"752149992a74e92d0b8d8ca0cbc09b450302b59ca2b8d75323dc518f2a4509c9"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114000051024430 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114000052017346 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113755732016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113755732015466 5ustar corecore